Home

A Course in Geophysical Image Processing with Seismic Unix

image

Contents

1. 14 14 24 24 i v 34 v 34 44 44 54 54 Shot at ep 200 no gain applied Shot at ep 200 tpow 1 offset meters x104 d offset meters x104 C 0 1 130 1 1382 1 1384 1 136 1 138 1 140 6 1 130 1 132 1 134 1 136 1 138 1 140 24 2 O O o 34 o3 E 4 4 54 54 REIR its BHi HRH I o Shot at ep 200 tpow 2 Shot at ep 200 jon 1 Figure 9 3 Gaining tests a no gain applied b tpow 1 c tpow 2 d jon 1 Note that in the text we often use jon 1 because it is convenient not because it is optimal It is up to you to find better values of the gaining parameters Once you have found those you should continue using those 121 AGC takes the rms amplitude of a seismic trace in a succession of windows on each seismic trace sums over the RMS va
2. Stolt simple su unfiltered trace number Stolt simple su filtered trace number 50 100 150 Stolt simple su interpolated Figure 7 3 a simple su data unfiltered b simple su data filtered with a 5 10 20 25 Hz trapezoidal filter c Stolt migration of unfiltered data d Stolt migration of filtered data e interpolated data f Stolt migration of interpolated data Clearly the most satisfying result is obtained by migrating the interpolated data a b 500 1000 1 500 2000 2500 3000 500 1000 1 500 2000 2500 3000 2 n 500 1000 1500 2000 2500 3000 d 500 1000 1500 2000 2500 3000
3. time section reconstructed Figure 5 2 a Test pattern b Test pattern corrected from time to depth c Test pattern corrected back from depth to time section Note that the curvature seen depth section indicates a non piecewise constant v t Note that the reconstructed time section has waveforms that are distorted by repeated sinc interpolation The sinc interpolation applied in the depth to time calculation has not had an anti alias filter applied 62 b velocity m s 0 2000 3000 4000 5000 6000 ENE f f 1 a 1500 2000 3000 M s velocity Pr 112 5 1000 4 Depth m 162 5 1500 4 Well Log c 1500 2000 3000 M S velocity 131 25 193 75 Figure 5 3 a Cartoon showing an idealized well log b Plot of a real well log A real well log is not well represented by piecewise constant layers c The third plot is a linearly interpolated velocity profile following the example in the text This approximation is a better first order approximation of a real well log 63 try to figure out the depths Z1 Z2 and Z3 necessary to undo the operation suztot lt junkl su z Z1 Z2 Z3 v 1500 2000 3000 gt junk2 su suxwigb lt junk2 su title time section reconstructed amp Please note you don t literally type z Z1 Z2 Z3 what you want is to find three numbers representing depths to substitute in for Z1 Z2 and Z3 The first value 71 0 You will notice t
4. 0 54 0 54 0 54 0 54 slopes 6 5 0 5 6 d slopes 4 3 0 3 4 C io 0 1000 2000 3000 de 0 1000 2000 3000 0 54 0 54 a r p 0 0 aie ot Fs ee a ore cs a 0 54 0 54 slopes 3 2 0 2 3 slopes 2 1 0 1 2 Figure 7 5 The k1 k2 domain plots of the simple su data with the respective dip filters applied in the Stolt migrations of Figure 7 4 102 suxwigb xcur 3 title Mmigration after dipfilter d2 40 amp sudipfilt dt 1 dx 1 lt simple su slopes 5 4 0 4 5 amps 0 1 1 1 0 sustolt cdpmin 1 cdpmax 80 dxcdp 40 vmig 2000 tmig 0 suxwigb xcur 3 title migration after dipfilter d2 40 amp sudipfilt dt 1 dx 1 lt simple su slopes 6 5 0 5 6 amps 0 1 1 1 0 sustolt cdpmin 1 cdpmax 80 dxcdp 40 vmig 2000 tmig 0 suxwigb xcur 3 title Mmigration after dipfilter d2 40 amp Dip filtering is less satisfying in this case because where it works well to eliminate the spatial aliasing it also eliminates the dips in the image which constitute the most steeply dipping parts of the structure The effect in the k1 k2 domain may be seen in Figure 7 5 The resulting migrations of the dip filtered versions of may be seen in Figure 7 4 Where the spatial aliased noise is suppressed the best the steeply dipping parts of the syncline are not imaged at all Again we see that trace interpolation is the best option for suppressing spatial aliasing 7 12 Concluding Remarks I
5. 98 The fan like shape results because we may identify k vectors with ray vectors The angualar range in the k domain represents that angular range of rays which illuminated the reflector in the simple model 7 11 4 Remedies for spatial aliasing Fundamentally spatial aliasing can only be completely avoided if data are sampled suff ciently finely in space to accurately represent all spatial frequencies alternately wavenum bers in the data As we have seen above simply having receivers more closely spaced significantly reduced the spatial aliasing in our test example While collecting the data with fine enough spatial sampling in the first place is the best remedy we may not alwatys have adequate spacing for all frequencies in the data Mathematically we can see how frequency bandwidth corresponds to spatial coverage If we write the wave equation 2 1 0 lv ES a u a t 0 and assume a solution of the form u x t Ae Een that is we assume a space and time oscillating solution then we obtain x i k wt a a VA z V A z le i k wt Canceling common terms and saving only the term with the highest powers in w and k we have k 0 H E implying that w w k ae Thus when w and k are large we have a relationship between k and w c a that tells us that reducing the bandwidth in w will also reduce the bandwidth in k Thus frequency filtering can be us
6. ZOD u6 MA sd u TPL sqof u d js x DIMO x S B YD pueunuop JO1 U0D SS99D01q dapiduioo JvIsvg oe fo od sapiduiod UVAJA SIL O LO S4OMA LOf APOD J YII ful 4 pduo2 J g fo 2 uondrns q puewuop J j dwo9 fJo pu mdmo sm Sfo 8umu8 q mdmo fpe saouasaffip f SIST Y If WP ud yomu mpy souy smdmo Sf ud dais Saf adowad ajajaq fu g sp f opf aumuay af fru sagaid auy u omu f mdg u mds fuos amp pon qoydyy fuos f on pf apf dop g If do Sof om aapdwuod gf Lf duis Jo apou uoysajosd adunyy Jf apou powyo Som Zp f sapuaypouoy S lt gY if eo uaasos q SJuaqUod Af ISTT Jf aow apyf fo squaquod 4S1 Jf wo JUNOD ADYD Y Psom UIT f OM AONPA 1X21 HI pe AONIPA UBAAISTN SIMU Lf soews 4Onpa UIIAISIMF IA LH ta uod nsq pueuwuop uonejndiue al 4 Piom suuo 4of Kajua qpnuvui XINN jipul fo uoyporfnou JUD_ISUT 4 SN 04 JDU Puas jouluuat BUDS dos J O1 UOISSAS JDUIULIA aADS dajuisd auy o1 f opf mdmo uondr sq ue awmu AUDU WEU us q aUIDUasn reu uolssas xo HJ duos f 4amuud q 1d pueuuop djaH 2 uonesiunwwo nding U pUDUUWOJ Juadad MUNG uj spupunluos juaras ASIG Kosy A10J90AIP BUDJAOM JUL pad uonpunofur aSU asupyy uyo uonpunofur sasn mdmo 2umusn su 4asn PUANI KANIA TUWLOYM SAISN Ul PASSO ISTT oym AU X AWP Ud ep ponb ysip amp ods q ejonb SANIDA JUAUUOMAUA JU saspyp puputuioo Kojds q aupu ausjutd awupu sere qiojap ut sapif Isr
7. 5 4 4 62 ghd a Soke ee ee ee 171 11 3 Lab activity 20 Predictive deconvolution 172 11 3 1 Spiking or ungapped Decon aoaaa aie a eb Ga 172 11 3 2 Multiple suppression by Wiener filtering gapped prediction error HIGEN Eene a s arei apa mae asi A Aer te ia a a a a r whos 175 11 4 What else did predictive decon do to our data 00 176 11 4 1 Deconvolution in the Radon domain 177 TEX ECON 444 3 3 3 Sige A le a She ee a OS ee 177 11 6 Lab Activity 20 Wavelet shaping 4 177 11 7 Advanced gaining operations 05144 Gh bo ae 4s le oes ee SG 179 11 7 1 Correcting for differing source strengths 180 11 7 2 Correcting for differing receiver gains 180 11 8 Muting NMO corrected data ds Ske BS eo BE SO lee Ges 181 11 9 Surface related multiple elimination 181 11 10Homework Assignment 8 Due Thursday 18 Oct 2012 before 9 00am 182 11 10 1 How are we doing on multiple suppression and NMO Stack 183 11 11Concluding Remarks aoaaa oho iy al tied ote io Ay hens Ba ke bo Boks tek 184 12 Velocity Analysis on more CDP gathers and Dip Move Out 185 12 1 Smarter velocity analysis using multiple suppression in the Radon Trans formi domain k ra hed oe a bea at OSL Gh ee ah S A in a se Se 187 122 NMO and Stack ao oe ee Se a le Bae A od ele Sed Bae 188 12 2 1 Homework 9 Velocity analysis for stack Due 25 Oct 2012 bef
8. You may want to compare these images with the Stolt migrations Ask the same questions as before Is the image better in any way Are there more artifacts or fewer artifacts Has your opinion about the background wavespeed changed 8 2 Lab Activity 10 FD FFD PSPI Split step Gaussian Beam v x z migrations To create better images having a migration that uses a background velocity profile that varies both in horizontal and vertical position u x z is required In the directory data cwpscratch Data4 we have several shell scripts that run 5 dif ferent types of migration These are finite difference Fourier finite difference phaseshift plus interpolation split step and Gaussian Beam GB All of these were discussed in previous sections with the exception of GB The GB stands for Gaussian beam and refers to the beam like approximate wavefield This choice of Green s function is often found to be an improvement on standard ray meth ods Gaussian beam migration was invented by geophysicist Ross Hill in about 1990 Typically people in the industry usually make a distinction between ordinary Kirchhoff migration and Gaussian Beam migration however Gaussian Beam migration is an ap plication of the Kirchhoff migration technique using Green s functions that are beams with an amplitude that varies like a bell shaped or Gaussian function in a cross section of the beam We are going to migrate the same data seismic3
9. first sample location for non seismic data float d2 sample spacing between traces float 2 first trace location float ungpow negative of power used for dynamic range compression float unscale reciprocal of scaling factor to normalize range int ntr number of traces short mark mark selected traces short shortpad alignment padding short unass 14 unassigned NOTE last entry causes a break in the word alignment if we REALLY want to maintain 240 bytes the following entry should be an odd number of short UINT2 OR do the insertion above the mark keyword entry float data SU_NFLTS segy Not all of these header fields get used all of the time Some headers are more important than others The most relevant fields to normal SU usage are the header fields tracl tracr dt cdp offset sx gx sy gy and delrt To see the header field ranges on sonar su radar su and seismic su type surange lt sonar su 59 584 traces tracl cdp muts ns dt 1 584 1 584 1 584 1 584 75 3000 100 surange lt radar su 501 traces trac 1 501 1 501 tracr 1 501 1 501 trid 1 ns 463 dt 800 hour 11 minute 3 33 3 33 sec 0 59 41 7 surange lt seismc su 801 traces tracl 1200 2000 1200 2000 tracr 67441 115081 67441 115081 fldr 594 991 594 991 tracf 18 2 2 ep 700 1100 700 1100 cdp 1200 2000
10. gt bZ l 1 The cross correlation of A Z and B Z is then the product of the polynomials A Z B Z bs az n The effect is to flip the order of the B z series in the multiplication to the opposite order as would be done with convolution 11 3 Lab activity 20 Predictive deconvolution There is a class of deconvolutional processes known as Wiener filters or prediction error filters which have been found to be useful in exploration seismic methods The method is called predictive because it assumes that the data have a specific character that allow later parts of the data to be related to earlier parts of the data Wiener filtering assumes that the data are minimum phase While there is a re quirement that the spectrum of the data is white a small noise parameter is added or assumed in the algorithm to prevent division by zero Physically if a waveform is minimum phase its energy is located in the front part of the waveform 11 3 1 Spiking or ungapped Decon The Wiener filter has the property that it calculates the inverse of an input waveform based on the delay or lag of the filter The inverse is not an exact inverse but rather a least squares estimate of the inverse If the minimum time delay of the filter the minimum lag is zero or at the very least the time sampling interval in the data and the length of the filter the maximum lag is chosen to be the autocorrelation length of the input data then t
11. time offset pairs that are zeroed out by sunmo and are thus missing after the cascade of processes that finishes with an inverse NMO 10 3 1 The stretch mute When we apply the normal moveout NMO correction there is a radical distortion of the data called the NMO stretch To remedy the NMO stretch an editing of the data called the stretch mute is applied You have already applied the stretch mute but you were unaware of this If you look at the self doc for sunmo sunmo SUNMO NMO for an arbitrary velocity function of time and CDP sunmo lt stdin gt stdout optional parameters Optional Parameters tnmo 0 NMO times corresponding to velocities in vnmo vnmo 1500 NMO velocities corresponding to times in tnmo cdp CDPs for which vnmo amp tnmo are specified see Notes smute 1 5 samples with NMO stretch exceeding smute are zeroed lmute 25 length in samples of linear ramp for stretch mute sscale 1 1 to divide output samples by NMO stretch factor invert 0 1 to perform approximate inverse NMO upward 0 1 to scan upward to find first sample to kill Notes For constant velocity NMO specify only one vnmo constant and omit tnmo 152 dial Ayala al tet kedik bih i elt diji ath acti Fak Eee r APEA KA Mea NEE Eb et S bide Latul MMJ hda alu it Miik na With ha ai Bub ait E AEE AA H rat ye hy Here ar thie Nia EN eter ein Bre bole array cure ethan aay Lea hd wu incitati de Udara ata ble HAL ke ara ea M
12. lt stack nmo radon gain jon 1 su gt stolt stack nmo radon gain jon 1 su 189 The file vintt bin is an approximate interval velocity as a function of time vj t and thus may be used as the vfile in sugazmig sumigps or suttoz For example sugazmig lt stack nmo radon gain jon 1 su vfile vintt bin dx 12 5 gt gazmig stack nmo radon gain jon 1 su sumigps lt stack nmo radon gain jon 1 su vfile vintt bin dx 12 5 gt migps stack nmo radon gain jon 1 su suttoz vfile vintt bin nz 1500 lt stolt stack nmo radon gain jon 1 su gt depth stolt stack nmo radon gain jon 1 su Warning At best these automated velocities are for testing purposes only Similarly there is file vintzav bin that may be used with the depth migration programs Again there tend to be systematic errors between stacking derived interval velocities and true interval velocities so these should be used for quick looks only 12 3 Homework Assignment 10 Preliminary Stolt migration Due 1 Nov 2012 before 9 00am Perform a Stolt migration of your multiple suppressed NMO corrected and stacked data Because Stolt migration uses RMS which is to say NMO stacking velocities take one of the tnmo and vnmo pairs from your nmovel par file as your tmig and vmig values Or make up what you view as a representative or average set of tmig and vmig values for sustolt For time ranges where you see smiles in the migrated data reduce the veloc
13. more MakeFake MakeFake 1s fakex The files that are generated all begin with the word fake suxwigb lt fake su perc 99 title fake data amp suxwigb lt faketwater su perc 99 title fake water bottom multiples amp suxwigb lt fake water tpegleg su perc 99 title fake water pegleg multiples amp Plots similar to these are shown in Figure 10 5a b and c Figure 10 5a shows a synthetic data panel similar to CDP 265 in the Viking Graben data without multiples The second panel Figure 10 5b shows the same data contami nated with simulated water bottom multiples Finally simulated water bottom multiples plus pegleg multiples from select events are shown in Figure 10 5c We can view semblance plots of each simulated datastes via suvelan nv 150 fv 1450 dv 15 lt fake su suximage d2 15 2 1450 verbose 1 title fake cmap hsv2 legend 1 bclip 5 amp suvelan nv 150 fv 1450 dv 15 lt faketwater su 142 1000 offset m 2000 1000 offset m 2000 3000 o nna ail IAA an ro inant a Synthetic data no multiples oO bza N a Synthetic data w b multiples Synthetic data w b pegleg multples Synthetic data plus simulated 265 b Synthetic data plus water bottom multiples Synthetic data similar to CDP Figure 10 5 a plus select pegleg gs n x E a z Q ae 82 eo S
14. 5 1 Imaging as the solution to an inverse problem Acoustic and elastic waves echo off of jumps in the wavespeed and or the density of the medium In the case of electromagnetic scattering the signal is coming from a volume of material or a layer rather than a boundary between layers which has a differing conductivity from the surrounding material In each case the propagating wave impinges on the reflector at some angle and is reflected from at an angle determined by the law of reflection for the medium For scalar waves which is to say waves that do not experience mode conversion the angle of incidence equals the angle of reflection For elastic waves the angle of reflection is a function of the angle of incidence and of the velocities and densities of the media on either side of the reflector The scattered wave therefore carries information about both the orientation of the reflector and its location Thus an image formed from such data is a solution to an inverse problem wherein the wavespeed of the medium and the location and orientation 59 x SP x gt Figure 5 1 Cartoon showing the simple shifting of time to depth The spatial coordinates x do not change in the transformation only the time scale t is stretched to the depth scale z Note that vertical relief looks greater in a depth section as compared with a time section of the reflector are the unknown variables being solved for 5 2 Inverse scattering imaging as time to
15. For example is not uncommon to want to look a the first N traces For example suwind key tracl count 1000 lt seismic su suximage perc 99 amp gives a quick look at the data We can see gathers of some variety To see what kind of gathers we have shot versus CMP the header values will help us Typing the following sugethw sx gx offset ep cdp lt seismic su more x 3237 gx 0 offset 3237 ep 101 cdp 1 x 3237 gx 25 offset 3212 ep 101 cdp 2 x 3237 gx 50 offset 3187 ep 101 cdp 3 x 3237 gx 75 offset 3162 ep 101 cdp 4 x 3237 gx 100 offset 3137 ep 101 cdp 5 x 3237 gx 125 offset 3112 ep 101 cdp 6 x 3237 gx 150 offset 3087 ep 101 cdp 7 x 3237 gx 175 offset 3062 ep 101 cdp 8 x 3237 gx 200 offset 3037 ep 101 cdp 9 x 3237 EX 225 offset 3012 ep 101 cdp 10 x 3237 gx 250 offset 2987 ep 101 cdp 11 x 3237 gx 275 offset 2962 ep 101 cdp 12 114 trace number 400 800 1000 600 first 1000 traces Figure 9 1 The first 1000 traces in the data 115 shows the values of several important header fields We can eventually figure out that these are shot gathers by noting which fields change the most slowly In this case the source position sx and the energy point number ep are the slowest changing Shot gathers are also called common shot gathers shot records and common source gathers These terms are used interchangeably It is a good idea to get to know your data by flipping through it much as you would
16. Seismic pressure sensor 12 Multicomponent seismic sensor Vertical component 13 Multicomponent seismic sensor Cross line component 48 14 Multicomponent seismic sensor in line component 15 Rotated multicomponent seismic sensor Vertical component 16 Rotated multicomponent seismic sensor Transverse component 17 Rotated multicomponent seismic sensor Radial component 18 Vibrator reaction mass 19 Vibrator baseplate 20 Vibrator estimated ground force 21 Vibrator reference 22 Time velocity pairs 23 N optional use maximum N 32 767 Following are CWP id flags 109 autocorrelation 110 Fourier transformed no packing xr 0 xi 0 xr N 1 xi N 1 111 Fourier transformed unpacked Nyquist xr 0 xiflO xr N 2 xi N 2 112 Fourier transformed packed Nyquist even N xr 0 xr N 2 xr 1 xi 1 xr N 2 1 xi N 2 1 note the exceptional second entry odd N xr 0 xr N 1 2 xr i xili xr N 1 2 1 xi N 1 2 1 xi N 1 2 note the exceptional second amp last entries 113 Complex signal in the time domain xr 0 xi 0 xr N 1 xi N 1 114 Fourier transformed amplitude phase a 0 p 0 a N 1 p N 1 115 Complex time signal amplitude phase a 0 p 0 a N 1 p N 1 116 Real part of complex trace from 0 to Nyquist 117 Imag part of complex trace from 0 to Nyquist 118 Amplitude of complex trace fro
17. amp exit 0 As before save ViewMig and change the mode via chmod u x ViewMig to make the file executable Run the command by typing ViewMig on the commandline ViewMig You may vary the parameters to change the appearance of the plot The idea of ViewMig is to give side by side comparisons of the various migration types 90 7 9 Homework 3 Due 13 Sept 2012 Rewrite the shell script Migtest combining ViewMig so that it saves plots as PostScript output i e use supswigb orsupsimage Compare the output of the different migra tion methods Take care to recognize that you may need vastly different values for the values of hbox and wbox by checking the self documentation for supsimage and psimage Take particular note of the appearance of the output any noise or artifacts you see Include comparison figures commands and commentary Again submit no more than 3 pages maximum as a PDF file 7 9 1 Hints PostScript is a graphics language created by Adobe systems which is the forerunner to PDF PostScript is still widely used and has not really been replaced by PDF On most systems there is a tool for viewing PostScript format files On Linux systems one such tool is gs which is the GhostScript interpreter GhostScript is a powerful graphics data conversion program Another PostScript viewer gv which is GhostView You may be able to view your eps files with gs gs filename eps You should also find that if you are using Op
18. division by zero is not defined We recall that w w is a complex valued function which may be written in complex exponential form as w w ww e or as the sum of real and imaginary parts as w w w w iw w We define the complex conjugate of w w as Blo fulo or as the sum of real and imaginary parts as w w wrw iwilw If we multiply w w by its complex conjugate we have the square of the modulus of w w as above I w w w w w w Returning to our deconvolution problem multiplying top and bottom of the integrand by w w we have S t Wx Dit Lp Wea iy 27 J w w We still haven t solved the problems of division by zero in w w because if w w has a zero then so will w w We solve this problem by adding a small number to the 170 denominator S t Wx Dit 1 re ww dw iw oa Leanne a The quantity is the noise or whitening or white noise parameter This parameter is chosen to be small enough to stabilize the inverse but not so big as to skew the results Thus formally we can define the inverse waveform W t by its Fourier transform representation 1 f w w Wee e dw 2r J co w w It is important to remember that no matter how a deconvolutional process is performed we think of deconvolution as division in the frquency domain All deconvolution schemes must then have the equivalent of a white noise parameter to stabilize the di
19. doy Mond dog Bae Wa SEK Mae ER os 8 1 3 Questions for discussion bes lt a ck panded AeGih dees wees He ae 8 2 Lab Activity 10 FD FFD PSPI Split step Gaussian Beam v z z METAIS a a 8 6 ord sais Bok Se GS ee ah ee Bt eee 8 3 Homework Assignment 4 Due 20 Sept 2012 Migration comparisons 8 4 Concluding Remarks uae sawed ae Ge Ee Eke Se 2s Data before stack 9 1 Lab Activity 11 Reading and Viewing Seismic Data 9 1 1 Reading the data syste erode at tay ake od bs Be eke Mek Be od 9 2 Getting to know our data trace header values 9 2 1 Setting geometry x snos stare ae Goo ce cerca suet eh ese S ater get 9 3 Getting to know our data Viewing the data 9 3 1 Windowing Seismic Data 22 5 4506 Bee te hod Oe eo 9 4 Getting to know your data Bad or missing shots traces or receivers 9 4 1 Viewing a specific Shot gather ec waves a 402 Ge a al ase es be 9 4 2 Charting source and receiver positions 9 5 Geometrical spreading aka divergence correction 9 5 1 Some theory of seismic amplitudes 9 5 2 Lab Activity 12 Gaining the data 0 50 Statisti al gaining o eA to oe a he te oe a aoe Bi eee oe ia 9 5 4 Model based divergence correction 204 9 6 Getting to know our data Different Sorting Geometries 9 6 1 Lab Activity 13 Common offset gathers 9 6 2 Lab Acti
20. flip through a coffee table picture book We can view all of the shot records by using suxmovie suwind count 12000 skip 0 lt seismic su suxmovie n2 1200 loop 1 perc 99 title Frame fg amp It takes a few minutes but eventually a window will come up which you can re size by dragging on the lower right corner This window will show a movie of 10 gathers at a time with the frame number being shown in the title You can stop the movie at a frame by pressing the far right mouse button You may see the successive 12000 trace blocks by setting skip 12000 skip 24000 and so forth Events with differing moveouts Stop the movie at any frame and zoom into view features of the shot gathers Some features to look for are multiples These are repetitions in the data in time caused by reverberations in the water column Pegleg multiples may appear to be arrivals with hyperbolic moveout that show a long time moveout in a gather Whereas reflections will have less moveout indicating higher velocity but also be hyperbolic in shape Reflections that have an hyperbola that peaks away from the shot position will indicate a dipping bed Direct arrivals will tend to have a linear moveout as will ground roll on land data that appears to roll over within the section 9 4 Getting to know your data Bad or missing shots traces or receivers In real data there may be bad traces or missing traces Some shots may be bad or there may be consistent o
21. gathers This is done in SU via susort offset gx lt seismic su gt seismic co su sugain jon 1 lt seismic co su gt gain jon 1 co su Again the choice of jon 1 for the gaining is used here for convenience You should use your own values of tpbow gpow and qclip instead File naming convention Note that the file names are chosen to reflect the processing steps applied to the data in the file gain jon 1 co su indicates that the file contains common offset gathers that have been gained with parameter jon 1 The convention is not unique but is convenient as it is easy to forget what processes have been applied to a file 9 6 2 Lab Activity 14 CMP CDP Gathers In 1950 geophysicist Harry Mayne patented the Common Depth Point Stacking method of seismic data enhancement The idea is simple sort the data into gathers whose source and receiver geometry all have the same midpoint in each gather Correct for the normal moveout NMO and sum stack The result should be less noisy equivalent zero offset trace To sort the data we use susort a cleverly written program that makes use of the powerful sorting capability built into the Unix operating system 9 6 3 Sort and gain Rather than make a bunch of intermediate temporary files we run the process of gaining and sorting successively in a pipe Our processing flow for gaining the full dataset is 123 midpoint meters 0 4 0 6 0 8 x104 b midpoint meters x104 0 2 0 4 0
22. seismic migration is viewed by many as the solution to an inverse problem wherein recorded seismic data are used as input to solve for the reflectivity of the reflectors as well as other important material parameters that characterize lithology 79 Chapter 7 Lab Activity 6 Several types of migration In this assignment we will apply several types of migration to the simple su data These types of migration represent many of those that are commonly used in industry 7 1 Different types of velocity All seismic migrations require a background wavespeed velocity profile However we must be very careful when addressing the term velocity The actual wavespeeds in the subsurface are called interval velocities and most likely are going to be a function of position However the types of velocities that we often encounter such as those obtained from velocity analysis are stacking velocity also known as the NMO velocity which is approximately the RMS root mean squared velocity Such velocities are often expressed as velocity as a function of time Having a velocity that is a function of time may seem strange at first but imagine that you have a seismic section with several strong seismic horizons These strong arrivals must represent relatively large impedence contrasts If we had a set of well logs to go with the seismic data then we could identify those rock units with the strong impedence contrasts We could bu
23. sstat Source static correction in milliseconds gstat Group static correction in milliseconds tstat Total static applied in milliseconds Zero if no static has been applied short laga Lag time A time in ms between end of 240 byte trace identification header and time break positive if time break occurs after end of header time break is defined as the initiation pulse which maybe recorded on an auxiliary trace or as otherwise specified by the recording system short lagb lag time B time in ms between the time break and the initiation time of the energy source may be positive or negative short delrt delay recording time time in ms between initiation time of energy source and time when recording of data samples begins for deep water work if recording does not start at zero time short muts mute time start 52 short mute mute time end unsigned short ns number of samples in this trace unsigned short dt sample interval in micro seconds short gain gain type of field instruments code 1 fixed 2 binary 3 floating point 4 N optional use short igc instrument gain constant short igi instrument early or initial gain short corr correlated 1 no 2 yes short sfs sweep frequency at start short sfe sweep frequency at end short slen sweep length in ms short s
24. suwind itmax 64 lt junki su suxwigb title time to depth amp where suwind has been used to pass only the first 64 samples of each trace The short answer is that while the time to depth and depth to time conversions are ostensibly simply piecewise linear operations in this simple example the fact is that there is the potential for errors to be introduced by the interpolation process These errors may make process of stretching only partially invertible The main problem is deciding how the v t and v z functions are to be interpolated As is seen in Fig 5 3 constant step models seen in cartoon well log diagrams does not accurately depict the complexity of actual well logs The simplest approximation to a real well log is a piecewise continuous curve with piecewise linear being the simplest curve That is we assume a functional form of v t mt b for velocity as a function of time For the example of t 0 0 15 2 v 1500 2000 3000 in the region from t 0 0 to t 15 the velocity profile would be given by v t 500 15 t 1500 and in the region from t 15 to t 2 the velocity profile is given by v2 t 1000 05 t 2000 Calculating the values of the depths 71 72 73 we see trivially that Z1 0 and that integrating the two equations above yields 72 131 25 and Z3 193 75 respectively 65 5 5 Sonar and Radar bad header values and incomplete information Most likely depth conversion for sonar and radar requi
25. t want The items to mute consist of random noise that may appear on the traces before the onset of the actual arrivals direct arrivals from the source that have traveled in the water layer refracted arrivals also called head waves from the shallower layers of the water bottom and wide angle reflections from the shallow layers of the water bottom that may be seen at longer offsets These items are undesirable because they do not fit the traveltime moveout of reflectors or do not fit with the theory of seismic reflection that we assume when migrating the data 10 3 3 Lab Activity 16 muting the data Muting is a simple process we define a curve in the data such that for those times and positions all values at earlier time are simply set to zero 10 3 4 Identifying waves to be muted On any part of the data where the waterbottom has more or less the same sort we can create an average of all shot profiles in that area In the case of these data the entire dataset has a fairly flat waterbottom We can resort the data so that it is in increasing offset with the traces of the same offset side by side susort dt offset lt seismic su gt junkl su where we have chosen dt as the first parameter because it is a header field that is the same for every trace We then can stack the data sustack key offset lt junk1 su gt supershot su so that each resulting trace is the average of all of the trace at that offset We may view the su
26. that you can supply input The proper input for a commandline is an executable file which may be a compiled program or a Unix shell script The command prompt is saying Type program name here Try running this command with and without the ampersand amp If you run suplane suxwigb The plot comes up but you have to kill the plot window before you can get your com mandline back whereas suplane suxwigb amp allows you to have the plot on the screen and have the commandline To make the plot better we may add some axis labeling suplane suxwigb title suplane test pattern labell time s label2 trace number amp Here the command is broken across a line so it will fit this page of this book On your screen it would be typed as one long line 24 time s trace number 10 20 30 0 05 0 10 0 15 0 20 0 25 suplane test pattern Figure 2 1 The suplane test pattern 25 to see a test pattern consisting of three intersecting lines in the form of seismic traces The data consist of seismic traces with only single values that are nonzero This is variable area display in which each place where the trace is positive valued is shaded black See Figure 2 1 Equivalently you should see the same output by typing suplane gt junk su suxwigb lt junk su title suplane test pattern labeli time s label2 trace number amp Finally we often need to have graphical output that ca
27. velconv lt infile gt outfile intype outtype optional parameters Required Parameters intype input data type see valid types below outtype output data type see valid types below Valid types for input and output data are vintt interval velocity as a function of time vrmst RMS velocity as a function of time vintz velocity as a function of depth zt depth as a function of time tz time as a function of depth Optional Parameters nt all number of time samples dt 1 0 time sampling interval ft 0 0 first time nz all number of depth samples dz 1 0 depth sampling interval fz 0 0 first depth nx all number of traces Example intype vintz outtype vrmst converts an interval velocity function of depth to an RMS velocity function of time Notes nt dt and ft are used only for input and output functions 81 of time you need specify these only for vintt vrmst orzt Likewise nz dz and fz are used only for input and output functions of depth The input and output data formats are C style binary floats 7 2 Stolt or f k migration To migrate the simple data in the computer we first begin with Stolt migration Stolt s method published in 1978 is a migration by Fourier transform and is often called f k migration If we consider migration to be a shifting of data which is expressed as a signal processing technique then we may consider that shifting to be done as a filtering process The data are sh
28. 03 1971 5 2404 53 tnmo 0 0 0 751065 0 998251 1 18839 1 40706 1 81587 2 47186 vnmo 1500 1608 31 1720 06 1808 53 1859 75 2008 75 2399 88 tnmo 0 0 0 713037 0 988744 1 32149 1 65424 1 90143 2 89017 vnmo 1500 25 1626 94 1696 78 1817 84 1887 69 2050 66 2623 38 tnmo 0 0 0 665501 1 01727 1 38804 1 70178 2 02502 2 93771 vnmo 1500 1594 34 1724 72 1850 44 1915 62 2157 75 2576 81 tnmo 0 0 0 636979 1 04579 1 41657 1 68277 1 87291 2 32925 3 53666 vnmo 1500 1575 72 1771 28 1817 84 1887 69 2069 28 2278 81 2777 03 tnmo 0 0 0 465851 0 789094 1 09332 1 4641 1 90143 2 43383 vnmo 1500 1538 47 1631 59 1780 59 1850 44 2022 72 2330 03 tnmo 0 0 0 675008 0 865151 1 20741 1 48312 1 92045 2 5289 3 13736 vnmo 1500 1585 03 1710 75 1771 28 1831 81 2041 34 2460 41 2823 59 tnmo 0 0 0 779587 1 14086 1 92045 2 2532 2 63348 3 3275 vnmo 1500 1640 91 1743 34 2027 38 2222 94 2371 94 2860 84 tnmo 0 0 0 646486 0 846137 1 15037 1 51164 1 82537 2 49087 2 98525 vnmo 1500 1575 72 1687 47 1817 84 1938 91 1980 81 2446 44 2735 12 188 tnmo 0 0 0 655994 0 874658 1 17889 1 57819 1 89192 2 47186 vnmo 1500 1575 72 1715 41 1817 84 1943 56 2008 75 2520 94 tnmo 0 0 0 598951 0 874658 1 17889 1 55917 1 9965 2 50038 3 80286 vnmo 1500 1585 03 1710 75 1827 16 1948 22 2078 59 2395 22 3135 56 tnmo 0 0 0 684515 0 931701 1 25494 1 56868 1 9965 2 1296 2 63348 vnmo 1500 1580 38 1738 69 1827 16 1920 28 2064 62 2153 09 2413 84 tnmo 0 0 0 789094 1 26445 1 64474 2 2532 3 00
29. 1 cdp su gained and deconvolved seismic data sorted in cdp s vpicks nmovel par output file of vnmo and tnmo values normpow 0 see selfdoc for suvelan slowness 0 see selfdoc for suvelan cdpfirst 1 minimum cdp value in data cdplast 2142 maximum cdp value in data cdpmin 128 minimum cdp value used in velocity analysis cdpmax 2110 maximum cdp value used in velocity analysis 185 dcdp 512 change in cdp for velocity scans fold 120 maximum number of traces per cdp gather dxcdp 12 5 distance between successive midpoints in full datas set The idea of these scripts is to take the full dataset window it into specific CDPs as suming an increment in CDPs allow the user to pick a semblance panel and view the resulting NMO corrected version of the data panel If the velocity picks are to the lik ing of the user then the script proceeds to the next CDP panel The end product is a collection of tnmo and vnmo values in a file called nmovel par for Velan and radonnmovel par for Velan radon The distinction is made because the velocities used for Radon transform based multiple suppression may not be quite the same as those used for stacking If we go in 512 cdp increments across the data we will sample the velocity approxi mately 4 times depending on the values of cdpmin and and cdpmax across the section This will give a representative collection of velocities that will be more repesen tative of the actual vel
30. 2raz 0 C which vanishes by Cauchy s theorem because the integrand is an analytic function of Z The contour integral in equation sifts through the each term of the series represen tation of S Z and returns the original sequence of values as discrete values giving us our original series of digital samples back The inverse Z transform is effectively the inverse Fourier transform as long as the contour C is the unit circle Z 1 168 Minimum phase in Z transforms The Z transform of our series given by Z is a polynomial of the N th degree whose zeros must be inside the unit circle given by Z 1 for the signal to be minimum phase This is often taken as a definition of minimum phase but it is difficult to see intuitively what this means or why this is important 11 2 2 Convolution with a wavelet Our digital data are convolved with a wavelet given by W t Dit W t S t T W r S t 7 dr Co ne 3 s w w w e dw 2m Joo Thus recorded data D t is the convolution of a wavelet W t with the reflectivity series S t The last line shows the Fourier transform domain form of convolution Convolu tion is multiplication in the frequency domain Convolution of Z transform representations In the language of Z transforms convolution of two signals is the multiplication of the two polynomial representations in Z of the functions 11 2 3 Deconvolution Deconvolution then is the inverse process which is t
31. 3 1 Viewing an SU data file Wiggle traces and Image plots Though we are assuming that the examples sonar su seismic su and radar su are finished products our mode of presentation of these datasets may change the way we view them entirely Proper presentation can enhance features we want to see suppress parts of the data that we are less interested in accentuate signal and suppress noise Improper presentation on the other hand can take turn the best images into something that is totally useless 3 1 1 Wiggle traces A common mode of presentation of seismic data is the wiggle trace Such a represen tation consists of representing the oscillations of the data as a graph of amplitude as a function of time with successive traces plotted side by side Amplitudes of one polarity usually positive are shaded black where as negative amplitudes are not shaded Be aware that such presentation introduces a bias in the way we view the data accentuating the positive amplitudes Furthermore wiggle traces may make dipping structures appear fatter than they actually are owing to the fact that a trace is a vertical slice through the data In SU we may view a wiggle trace display of data via the program suxwigb For example viewing the sonar su data as wiggle traces is done by redirecting in the data file into suxwigb suxwigb lt sonar su amp the ampersand amp means run in background 32 This should look horrible
32. 4 war yk ie eke ee ee T2 ROM OE migration o ea hud ou ae Boo etl See lie Se oh Bly 7 2 1 Stolt migration of the Simple model data 7 3 Gazdag or Phase shift migration 2 0 0 0 2 00 000 7 4 Claerbout s 15 degree finite difference migration 7 5 Ristow and Ruhl s Fourier finite difference migration 7 6 Stoffa s split step migration S02 tug s Aad et E Bok a 7 7 Gazdag s Phase shift Plus Interpolation migration 7 8 Lab Activity 7 Shell scripts o 2 5 6 wos a ase Ga woke Re RL Pe He 7 9 Homework 3 Due 13 Sept 2012 ice a ol eta ak amp ate ee eh pe kiew amp HOGA CRON aa ea re Rae Hh yet a Ad Bw Nhe AIS HES Ze lake th 7 10 Lab Activity 8 Kirchhoff Migration of Zero offset data TIT Spatialkalasihg i ma bai els ct Beg oh ee 8 Begs bh hen ten ay hea D inde died 7 11 1 Interpreting the result 4 4 amp swt Ke oe amp A BAR a8 AS 7 11 2 Recognizing spatial aliasing of data in the space time domain 7 11 3 Recognizing spatial aliasing in the f k domain 7 11 4 Remedies for spatial aliasing 0 2 7 12 Concluding CMa Ke edt vee aS ae gE ak ig eo keg cle Zero offset v t and v x z migration of real data Lab Activity 9 8 1 Stolt and Phaseshift v t migrations 46x00 sh oka we Powe see x 8 1 1 Questions for discussion Ac 2 74 8 ede Hd ete Qe hog amp S 8 1 2 Phase Shift migration lt
33. 6 0 8 1 0 1 2 time s common offset 262 meters common offset 1012 meters c midpoint meters x104 0 2 0 4 0 6 0 8 1 0 1 2 time s common offset 3237 meters Figure 9 4 Common Offset Sections a offset 262 meters b offset 1012 meters c offset 3237 meters Gaining is done via sugain jon 1 for convenience A better gaining of the data is possible 124 susort cdp offset lt seismic su gt seis cdp su sugain jon 1 lt seis cdp su gt gain jon 1 cdp su though here jon 1 should be replaced with the best values for the gaining parameters that you can find File naming convention Note again that the file names are chosen to reflect the processing steps applied to the data in the file gain jon 1 cdp su indicates that the file contains common depthpoint gathers that have been gained with parameter jon 1 The convention is not unique but is convenient as it is easy to forget what processes have been applied to a given data file If you are experimenting with gains sort first because this is a much more expensive operation susort cdp offset lt seismic su gt seismic cdp su and then do the gain on the sorted data seismic cdp su sugain YOUR GAIN PARAMETERS HERE lt seismic cdp su gt gain PARAMETERS cdp su You will note that we always sort from file to file We have found that some systems the sorting will fail if susort is used with a pipe Again note the naming co
34. PATH will yield lib u yourusername bin usr bin X11 usr local bin bin usr bin usr local bin usr sbin usr local cwp bin 1 12 2 The SH family The process is similar for the SH family of shells The file of interest has a name of the form profile bashre and the bash_profile The bash_profile is read once by the shell but the bashrc file is read everytime a window is opened or a shell is invoked Or vice versa depending on the system Mac OS X seems to have a strange convention Thus what is set here influences the users complete environment The default form of this file may show a path line similar to PATH PATH HOME bin usr local bin which should be edited to read export CWPROOT usr local cwp PATH PATH HOME bin usr local bin CWPROOT bin The important part of the path is to add the CWPROOT bin on the end of the PATH line no matter what it says The user then logs out and logs back in for the changes to take effect In each case the PATH and CWPROOT variables are necessary to be set for the users working shell environment to find the executables of Seismic Unix 1 13 Unix help mechanism Unix man pages Every program on a Unix or Unix like system has a system manual page called a man page that gives a terse description of its usage For example type man ls man cd man df man sh man bash man csh to see what the system says about these
35. Prepare a report of your results The report should consist of Your plots a short paragraph describing what you saw Think of it as a figure caption a listing of the actual commandlines that you ran to get the plots Not more than 3 pages total Make sure that your name the due date and the assignment number are at the top of the first page e Save your report in the form of a PDF file and email to john dix mines edu 3 6 Concluding Remarks There are many ways of presenting data Two of the most important questions that a scientist can ask when seeing a plot are What is the meaning of the colorscale or grayscale of a plot and What normalization or balancing has been applied to the data before the plot The answers to these questions may be as important as the answer to the question What processing has been applied to these data 3 6 1 What do the numbers mean The scale divisions seen on the plots in this chapter that have been obtained by running suximage with legend 1 show numerical values values that are changed when we apply display gain Ultimately these numbers relate to the voltage recorded from a transducer a geophone hydrophone or accelerometer While in theory we should be able to extract information about the size of the ground displacement in say micrometers or the pressure field strength in say megapascals there is little reason to do this Owing to detector and source couplin
36. Radon domain 140 trace number 20 40 60 80 100 120 time s Steeply dipping plane removed Figure 10 4 The suplane test pattern data with the steepest dipping arrival surgically removed in the Radon domain 141 10 1 1 How filtering in the Radon domain differs from f k filtering But so what We suppressed a dipping arrival Couldn t we have done that with f k filtering For example the program sudipfilt could be used suppress the same arrival The Radon transform need not be applied only to straight lines which represent a range of dips in the data Thus it is possible to perform the Radon transform by passing more general curves through the data such as the curves we obtain by performing the NMO correction on CMP gathers We can attack the differing moveouts between reflections and multiples 10 1 2 Semblance and Radon for a CDP gather Our principal sources of multiples come from two sources The first are simple water bottom multiples which are reverberations in the water layer The second are called peg leg multiples an allusion to the ray path in the water layer resembling a crude artificial leg which are reflected arrivals that have one or more additional bounces in the water layer A crude simulation of our CDP 265 may be made with the MakeFake shell script cd scratch yourusername mkdir Temp5 if you have not done so already cp data cwpscratch Data5 MakeFake scratch yourusername Temp5
37. The problem is that there are 584 wiggle traces side by side Place the cursor on the plot and drag while holding down the index finger mouse button This is called a rubberband box Try grabbing a strip of the data of width less than 100 traces by placing the cursor at the top line of the plot and holding the index finger mouse button while dragging to the lower right Zooming in this fashion will show wiggles The less on here is that you need a relatively low density of data on your print medium for wiggle traces Place the mouse cursor on the plot and type q to kill the window Try the seismic su and the radar su data as wiggle traces via suxwigb lt seismic su amp suxwigb lt radar su amp In each case zoom in on the data until you are able to see the oscillations of the data 3 1 2 Image plots The seismic data may be thought of as an array of floating point numerical values each representing a seismic amplitude at a specific t x location A plot consisting of an array of gray or color dots with each gray level or color representing the respective value is called an image plot If we view An alternative is an image plot suximage lt sonar su amp This should look better We usually use image plots for datasets of more than 50 traces We use wiggle traces for smaller datasets 3 2 Greyscale There are only 256 shades of gray available in this plot If a single point in the dataset makes a
38. a8 vnmo 1500 CNMO 0 55 pec e gia tg ack gee vnmo 1500 5 where the tnmo vnmo are those that you got for the respective CDPs Note that the number of tnamo and vnmo values per pair have to be the same but each pair may have a different number of values from other pairs Use this radonnmovel par for performing the NMO correction before stacking sunmo par radonnmovel par lt radon gain yourgainparameters cdp su sustack gt stack su Obviously you cannot complete this assignment on time if you do not start working on this immediately Everyone must have a Radon multiple suppressed version of the full dataset radon gain yourgainparameters cdp su ready for class on the 28th so we can proceed with velocity analysis Other tips e Feel free to use at to run the jobs at night e Furthermore you might consider regaining the data after you have done multiple suppression e Who says that you need to stack all of your data It may be that the far offsets and the nearest near offsets could be omitted from your data and make the dataset a bit smaller 11 10 1 How are we doing on multiple suppression and NMO Stack The subset approach that is pursued in the Homework 7 and 8 suffers from a serious flaw While we have a set of stacking velocities for each block we are not taking advantage of the ability of our programs to interpolate these values across the section We may see a blocky appearance While pe
39. and some estimate of ray theoretic amplitudes need to be com puted for the background wavespeed profile to do the migration If you recall we might view migration as moving data up or down along a light cone Approximating the wave field by ray tubes is one way of doing that The shell script Rayt2d simple runs the program rayt2d which generates the nec essary traveltime tables The shell script Kdmig2d simple runs sukdmig2d on the simple su data Read each shell script by typing more Rayt2d simple bin sh rayt2d vfile vel kdmig simple dt 004 nt 501 fa 80 na 80 fz 0 nz 501 dz 4 fx 0 nx 80 dx 40 ek 0 nxs 80 fxs 0 dxs 40 ms 1 tfile tfile simple xmovie lt tfile simple clip 3 n1 501 n2 80 loop 1 title g traveltime table amp You may type q to get out of the more program The program rayt2d generates a traveltime table containing the traveltime from each source to each receiver for every point in the model The movie shown by xmovie shows these traveltimes as shades of gray The idea of running the movie over the traveltimes is to see if there are any inconsistencies in the collection of traveltimes more Kdmig2d simple bin sh sukdmig2d infile simple su outfile kdmig simple su ttfile tfile simple ft 0 fzt 0 nzt 501 dzt 4 00 angmax 80 0 fxt 0 nxt 80 dxt 40 fs 0 ns 80 ds 40 dxm 40 v0 2000 noff 1 off0 0 doff 0 You may type q to get out of the more program 93 The program sukdmig
40. applying to data 166 11 2 An overview of deconvolution mathematics Geophysicists have realized a tremendous benefit from a simple concept That is the concept called the convolutional model of seismic wave propagation Because the wave equation is a linear equation the simplest way of looking at all processes involving the wave equation is to consider the processes as being a linear system The common metaphor is that of a black box That is we are assuming that a geophysical process is a linear system with an input and an output but the only other thing that we know about the black box is that it behaves in a linear fashion That is we assume that the output from a process depends linearly on the input If multiply the input to a linear system by a scalar then the output is scaled by the same value If we shift the input to the black box the output is shifted by the same amount That s it Linear systems have a property called the principle of superposition which means that new solutions of a linear problem may be formed by the linear combination of other solutions that we already have in hand This allows transform theory to be applied and tells us through a result known as Green s theorem that if we have a particular solu tion called the Green s function of a linear system we may form all possible solutions by the convolution of a given input with the Green s function The Green s function is also known as
41. balances the data by shot gather panel The gaining of choice is to divide by the RMS power of the data which is the square root of the sum of the squares of the seismic data values in all of the traces of a given shot gather In a shell script we run rm pbal shot su for i in ls split_ do sugain panel 1 pbal 1 lt i gt gt pbal shot su done rm split The double redirect out gt gt says append values so the file pbal shot su contains all of the power balanced shots We are free remove the separate shot files after the process is complete via rm split_ 11 7 2 Correcting for differing receiver gains Similarly we can take the resulting power balanced shot gathers and sort these into receiver gathers via susort gx offset lt pbal shot su susplit key gx close 1 The result now is a collection of files whose names begin in the word split each containing a single receiver gather As before we run in a shell script the operations rm pbal rec su for i in ls split_ do sugain panel 1 pbal 1 lt i gt gt pbal rec su done rm split 180 to yield the power balanced receiver gathers The final file pbal rec su of receiver gathers can be re sorted into CDP gathers and gained and the other processing we have discussed already can begin susort lt pbal rec su cdp offset sugain jon 1 gt gain jon 1 cdp su where here we recognize that gain also includes power balancing for shot st
42. commands For example 22 man 1s LS 1 User Commands LS 1 NAME ls list directory contents SYNOPSIS ls OPTION FILE DESCRIPTION List information about the FILEs the current directory by default Sort entries alphabetically if none of cftuvSUX nor sort Mandatory arguments to long options are mandatory for short options too a all do not ignore entries starting with A almost all do not list implied and MORE The item at the bottom that says MORE indicates that the page continues To see the rest of the man page for Is is viewed by hitting the space bar View the Unix man page for each of the Unix commands you have used so far Most Unix commands have options such as the Is a which allowed you to see files beginning with dot or Is l1 which shows the long listing of programs Remember to view the Unix man pages of each new Unix command as it is presented 23 Chapter 2 Lab Activity 1 Getting started with Unix and SU Any program that has executable permissions and which appears on the users PATH may be run by simply typing its name on the commandline For example if you have set your path correctly you should be able to do the following suplane suxwigb amp this symbol the ampersand indicates that the program is being run in background the pipe symbol The commandline itself is the interactive prompt that the shell program is providing so
43. cwpscratch Data3 PSmig cp data cwpscratch Data3 seismic3 su cp data cwpscratch Data3 Suttoz stolt cp data cwpscratch Data3 Suttoz psmig AAA H HH These data were collected over a structure in the North Sea known as the Viking Graben The data were shot by Mobil Corporation back in the 1980s This is a stacked section The data seismic3 su have been gained have had multiple suppression applied via supef Wiener spiking and Wiener prediction error filter deconvolution have been NMO corrected velocity analysis performed with suvelan and normal moveout cor rected via sunmo dip moveout corrected with sudmofk and have finally been stacked using sustack The pre processing is not very good No migration has been applied First we check the header values on the data with surange surange lt seismic3 su 2142 traces tracl 1 2142 1 2142 tracr 1 120120 1 120120 fldr 3 1003 3 1003 tracf 1 120 1 120 ep 101 1112 101 1112 104 cdp 1 2142 1 2142 cdpt 1 120 1 120 trid 1 nhs 1 60 1 1 gelev 10 selev 6 scalel 1 scalco 1 SX 3237 28512 3237 28512 gx 0 28250 0 28250 counit 3 mute 48 ns 1500 dt 4000 It is convenient to copy and paste this screen into a file called Notes for future reference We see that there are 2142 traces in the section that the time sampling interval dt 4000 which is to say 4ms sampling and ns 1500 samples per trace We know from the he
44. depth conversion In geophysics there are three common types of inverse scattering imaging techniques that may be encountered These are sonar ground penetrating radar GPR and reflection seismic In each case a species of wave is introduced into the subsurface This wave is reflected off of structures within the Earth and travels back up to the surface of the Earth where it is recorded In the raw form the coordinates of the data consist of the spatial coordinates of the recording position and traveltime which may be represented as the ordered triple of numbers Data x1 2 t It is implied that some form of processing is needed to convert data collected in the input coordinates of space and time Data x 22 t into an image in the output coordinates that are purely spatial DepthImage y1 y2 y3 or are new spatial coordinates and a migrated time coordinate TimeImage y1 y2 T When the output is in space and migrated time we call the process time migration and the output a time section When the output is in purely spatial coordinates we call the process depth migration and the output a depth section Each type of section is found useful in exploration seismic 60 Thus for our migration as depth conversion we will consider the final step of pro cessing as a process that converts the data from Data y1 y2 T to data in DepthImage y1 ye y3 in purely spatial coordinates The simple
45. everything to the left of p 0 is data that we want to keep Water bottom multiples are flattened to p 0 and pegleg multiples fall somewhere between the data we want to keep and the water bottom multiples A more sophisticated of the NMO correction can be used as a preprocess to make the parts of the data we want to keep fall to the left of p 0 while moving items we want to remove to the right of p O 146 10 2 Multiple suppression Lab Activity 17 Radon transform As we may see in the synthetics in Figure 10 5 multiples tend to have steeper moveouts which is to say that the multiple energy takes longer time to travel the same distance because a leg of propagation has occurred in the water layer If we NMO correct our data to the water speed or maybe a speed that is slightly higher this will tend to flatten many of the multiples but cause events that we want to save to curve up We perform the NMO correction with sunmo sunmo vnmo 1500 lt gain jon 1 cdp 265 su suxwigb You may find that making your wiggle trace plot tall and narrow accentuates the different moveouts Anything that travels with the speed of water waves is flattened with this choice of NMO velocity Arrivals that have the moveout of reflections are now curving upward Anything horizontal near horizontal or curving down is something that we want to suppress We save a copy of this water speed NMO corrected data as sunmo vnmo 1500 lt gain jon 1 cdp 265 su gt jun
46. filtering For example we might consider simply applying a filter to the data as part of the processing sufilter f 0 2 60 70 where the values of the corner frequencies of the filter are chosen to reflect a reasonable range of frequencies in the data that can be trusted So finally the processing sequence for our fake data with waterbottom multiples is sunmo vnmo 1800 lt faketwater pegleg su supef maxlag 0673 supef minlag 068 maxlag 51 sunmo invert 1 vnmo 1800 sufilter f 0 2 60 70 suxwigb perc 99 xcur 2 For the real data some variation on this processing flow in terms of the values of min lag and maxlag will exist Indeed these values are guaranteed to vary across the survey 176 11 4 1 Deconvolution in the Radon domain Another possibility is to apply the prediction error filtering in the radon domain For example employing the linear T p transform we forward radon transform the data apply the prediction error filtering sunmo vnmo 1500 lt gain jon 1 cdp 265 su suradon choose 0 igopt 3 pmin 1500 pmax 1000 interoff 262 offref 3237 supef minlag 15 maxlag 1 0 suradon choose 4 igopt 3 pmin 1500 pmax 1000 interoff 262 offref 3237 sunmo vnmo 1500 invert 1 gt radonpef su The process will do a good job on simple water bottom reverberations but other multiples will not be as well suppressed unless these can be made exactly periodic in the radon domain 11 5 FX Decon There
47. increasingly larger angles have been created An im plementation of a version of this finite difference migration that gives the user a choice of different angles of accuracy is sumigfd SUMIGFD 45 90 degree Finite difference migration for zero offset data sumigfd lt infile gt outfile vfile optional parameters Required Parameters nz number of depth sapmles dz depth sampling interval vfile name of file containing velocities see Notes below concerning format of this file 85 Optional Parameters dt from header dt or 004 time sampling interval dx from header d2 or 1 0 midpoint sampling interval dip 45 65 79 80 87 89 90 Maximum angle of dip reflector Try cp data cwpscratch Data2 vel fdmig simple scratch yourusername Temp2 cd scratch yourusername Temp2 sumigfd lt simple su dx 40 dz 12 nz 150 vfile vel fdmig simple gt fd simple su Note the velocity file vfile expects interval velocities as a function of x z where zx is taken as the fast dimension in this file 7 5 Ristow and Ruhl s Fourier finite difference migration A hybridization between phase shift migration and finite difference migration known as Fourier finite difference was published in 1994 by D Ristow and T Ruhl This type of migration is implemented in sumigffd SUMIGFFD Fourier finite difference migration for zero offset data This method is a hybrid migration which combines the advantages of phase shift and finite
48. information Again simple is better than complicated and fewer depth horizons no more than 4 are better than more We then may use the script Horizon to pick depth horizons on the depth migrated image The shell script uses a similar blind picking technique as is used in Velan and prompts the user for velocities as well as the space rate of change of velocity these should be numbers on the order of dvdz 1 and dvdx 001 or dvdx 001 where the units are velocity units per meter The Horizon script is interactive and produces several output files Two of these are unif2 ascii and unif2 par These files is used by the shell script Unif2 sh to build a smoothed velocity file The files produced by Horizon may be hand edited and Unif2 run again to produce an updated set of velocity files The contents of Unif2 sh are bin sh set x set parameters here method mono interpolation method nz 1500 nx 2136 pickedvalues junk1 picks this is the output of Horizon cp junki picks temp ascii smoothint2 lt temp ascii r1 100 ninf ninf gt unif2 ascii 196 unif2 lt unif2 ascii par unif2 par smooth2 ni nz n2 nx r1 50 r2 50 gt junk bin cp junk bin unewvelzx bin transp ni nz lt unewvelzx bin gt unewvelxz bin exit 0 The value of nx is the number of CDPs in the stacked data The user may set the values of rl and r2 to higher or lower values to control the level of smoothing The output velocity files are unewv
49. mt mt suxmovie clip 1 0 title Acoustic Finite Differencing windowtitle Movie labeli label1 label2 label2 ni n1 d1 d1 f1 f1 n2 n2 d2 d2 f2 f2 loop 1 sleep 8 amp exit 0 You may run the demo by typing XSyncline The result shows the wavespeed profile for the model This is similar to the simple model that will be discussed later in the these notes A movie showing snapshots of the wavefield will begin Watch the wavefront of the energy from the shot expand You f2 may stop and restart the movie by pressing the far right mouse button Of interest are the frames at which the first reflections begin As the movie progresses you will see the reflected field progress as the reflection point propagates along the reflector surface Indeed from viewing this movie we can see why an integral over the reflector surface called the Kirchhoff modeling formula is a way of modeling the reflected field Note that you only see wavefronts there is nothing like a ray to be seen A ray is the trajectory taken by a point on a wavefront Second notice that the bowtie forms as the caustic in the propagating wavefield travels to the surface The movie will run in a loop You may stop the movie by pushing the right mouse button You may reverse the movie by pressing the middle mouse button Effectively running the field backward in time is reverse time migration In seismic data we do not have a reco
50. really suitable to be used for migration as is but may provide important information for con structing velocity models later on 186 12 1 Smarter velocity analysis using multiple suppression in the Radon Transform domain If we recall how we did multiple suppression it was by first performing an NMO correction to the water speed to flatten the multiples and then transforming the data into the Radon transform domain Using suradon with the option choose 1 allows the user to apply a filter defined by the line t 0 0 pmulb MIN VALUE t 6 0 pmula MAX VALUE where the values of pmula and pmulb are the intercepts at the beginning and ending times in the Radon transform domain We used pmula 0 and pmulb 0 for simplicity and used a preprocess of an NMO correction to change the moveout of the data to separate the presumed signal from the multiples by separating the moveouts If the multiples are at the water speed or slower these arrivals will have their moveouts changed to curve down while all good arrivals which is to say those containing non multiply scattered arrivals theoretically should curve up We can use the Velan radon script to improve this NMO preprocess by picking NMO velocities to use in the radon multiple suppression We then run the radon operation in a modified version of Radon full called Radon final bin sh data data gain jon 1 cdp su radondata radon data parfile nmovelradon par interoff 262 of fref 3237 de
51. see a lot of diffractions then this means the processing from noise suppression to velocity anal ysis to NMO and stack were done well enough to make the diffractions not be obliterated In this sense seeing a lot of diffractions after stack is good 110 Chapter 9 Data before stack So far all of our data sets have been zero offset synthetic data and poststack datasets We now seek to investigate the world of prestack data While we will find that prestack migrations are beyond the capability of the machines in the lab except for very small examples we will find that there are a host of prestack processes that we can apply Indeed most of the processing in seismic data processing is on prestack data We will find that we will be able to see far more in our dataset than we saw in the examples of the previous chapter Indeed students have the experience of making better images than published images on our dataset As we proceed students also may notice that we are having more fun because this is more like real industry processing unlike the first few chapters where our data were largely test patterns 9 1 Lab Activity 11 Reading and Viewing Seismic Data For the lab examples that follow make a temporary directory in your working area called Temp5 area and type cd scratch yourusername mkdir Temp5 As was discussed at the beginning of these notes one of the popular seismic data exchange formats is the SEG Y d
52. student is given a word picture and chalkboard introduction of the process of seismic data acquisition and the application of a myriad of processing steps for converting raw seismic data into a scientifically useful picture of the earth s subsurface This lab is designed to provide students with a practical hands on experience in the reality of applying seismic processing techniques to synthetic and real data The course however is not a training course in seismic processing as one might get in an industrial setting Rather than training a student to use a particular collection of software tools we believe that it is better that the student cultivate a broader understanding of the subject of seismic processing We seek also to help students develop some practical skills that will serve them in a general way even if they do not go into the field of oil and gas exploration and development Consequently we make use of freely available open source software the Seismic Unix package running on small scale hardware Linux based PCs Students are also encour aged to install the SU software on their own personal Linux or Mac PCs so that they may work and play with the data and with the codes at their leisure Given the limited scale of our available hardware and time our goal is modest to introduce students to seismic data processing through a 2D single component processing application The intended range of experience is approximatel
53. technique Part of this depends on what we want from the data Are amplitudes important or is it all image quality alone Which is more important speed or accuracy Time is money but then an image that fails to direct the driller to the correct target may be more costly in the long run Phase shift and finite difference algorithms are more expensive but may offer better images for the reason that the process may be thought of as wavefront continuation Maintaining the continuity of the backward propagating wavefront should in theory make a more physically correct imaging process Kirchhoff migrations are less computa tionally expensive but these depend on shooting rays or ray tubes Rays are sensitive to small variations within a model If the background wavespeed profile is not sufficiently smooth i e twice differentiable then inconsistencies in the calculated ray field may re sult with small variations introducing larger errors Thus any cost savings realized may be done at the expense of improper reconstruction or propagation of wavefronts Stolt migration relies on neither ray tracing nor wavefront continuation but on the assumption that a stretching and filtering process in the f amp domain can accurately undo the effects of wave propagation on the wavefield at the expense of never having the correct background wavespeed Stolt migration is fast Finally the beginning student learns that diffractions are bad However if we
54. that indicates the place where the user logs in The user is then prompted for a password Once on the system the user either has a windowed user 12 interface as the default or initiates such an interface with a command such as startx in Linux 1 4 What is a Shell Some of the difficult and confusing aspects of Unix and Unix like operating systems are encountered at the very beginning of using the system The first of these is the notion of a shell Unix is an hierarchical operating system that runs a program called the kernel that is is the heart of the operating system Everything else consists of programs that are run by the kernel and which give the user access to the kernel and thus to the hardware of the machine The program that allows the user interfaces with the computer is called the working shell The basic level of shell on all Unix systems is called sh the Bourne shell Under Linux based systems this shell is actually an open source rewritten version called bash the Bourne again shell but it has an alias that makes it appear to be the same as the sh that is found on all other Unix and Unix like systems The common working shell environment that a user is usually set up to login in under may be csh the C shell tesh the T shell which is a non proprietary version of csh ksh the Korn shell which is proprietary zsh which is an open source version of Korn shell or bash which is an open source version of the Bourn
55. the SEG Y format The first 240 bytes of each seismic trace in a SEG Y dataset consist of this trace header The data are always uniformly sampled in time so the data portion of the trace consisting of amplitude values only follows immediately after the trace header While it may be tempting to think of a seismic section as an array of traces in the computer these traces simply follow one after the other The part of the listing from sukeyword that is relevant at this point is skipping typedef struct segy trace identification header int tracl Trace sequence number within line 47 numbers continue to increase if the same line continues across multiple SEG Y files int tracr Trace sequence number within SEG Y file each file starts with trace sequence one int fldr Original field record number int tracf Trace number within original field record int ep energy source point number Used when more than one record occurs at the same effective surface location int cdp Ensemble number i e CDP CMP CRP int cdpt trace number within the ensemble each ensemble starts with trace number one short trid trace identification code 1 Other O Unknown Seismic data Dead Dummy Time break Uphole Sweep Timing Water break Near field gun signature OMAN ODOT PWN KE Il m e Il Far field gun signature 11
56. the times in the tO field These could be used for model building but are not used in migration Again these values are only preliminary If you look inside the shell script Suttoz stolt more Suttoz stolt you will see that the same velocities are used for depth conversion from the Stolt migrated data which is in time to make a Stolt migrated depth section Neither of these sets of velocities should be viewed as exact these are only prelim inary estimates Notice for example that there is no lateral variation in these velocities These are only v t which implies a v z rather than a v x z profile Yet a cursory examination of the data shows a profile that dips to the right indicating that there is likely substantial lateral variation of the velocities in the subsurface You may run the Stoltmig shell script by typing Stoltmig you will see that the output file is stolt seis su You may create a depth section version of stolt seis su by typing Suttoz stolt The resulting output is the file stolt depth seis su You may now plot all three of these files seismic3 su stolt seis su and stolt depth seis su via suximage lt seismic3 su clip 2 title Stacked data amp suximage lt stolt seis su clip 2 title Stolt time section amp suximage lt stolt depth seis su clip 2 title Stolt depth section amp 8 1 1 Questions for discussion Compare the original data to the Stolt time migration Co
57. to later times in the data we have to deal with the fact that the sample spacing of the time to depth shifted data changes as the velocity changes Indeed constant or piecewise constant profiles rarely accurately represent wavespeed variation in the real earth The depth is calculated for each sample but because we want the output to be uniformly sampled we have to interpolate the missing depth values This interpolation may be done many ways but in this program it is done by fitting a sinc function sinc interpolation to the data points Look up since interpolation in a textbook on signal processing The bandwith of this sinc function is the the band from 0 to the Nyquist frequency of the data When resampling to a greater number of samples the Nyquist frequency of the output is greater than the Nyquist frequency of the input so there is no possibility of aliasing However if we subsample data the potential for aliasing exists To repeat the test we should be setting nt 64 to force the number of samples to be the same on both input and output suplane gt junk su suttoz lt junk su t 0 0 15 2 v 1500 2000 3000 gt junki su suztot lt junki su nt 64 z Z1 Z2 Z3 v 1500 2000 3000 gt junk2 su suxwigb lt junk su title test pattern amp suxwigb lt junkl su title depth section amp suxwigb lt junk2 su title reconstructed time section amp AAA AHH The time to depth may be improved by truncating the additional values
58. top of the other producing a fat trace rather than a stacked trace We are talking 6 fold data primarily so there were a maximum of six traces plotted side by side very closely Thus Harry Mayne s patent encouraged the development of high resolution plotting equipment Prior to the mid 1970s deconvolution followed by the CDP NMO STACK sequence was the majority of seismic data processing Migration was done only on select sections or not at all Seismic interpreters extracted dip and depth information from these stacked data 134 Chapter 10 Velocity Analysis Preview of Semblance and noise suppression Here we do a dry run of velocity analysis on a single CMP gather and then do a more production level approach to velocity analysis in the next chapter The method of velocity analysis that we will use is called NMO Semblance analysis The idea is to apply the normal moveout correction over a spectrum of velocities and pick the velocities that NMO correct the data to be most coherent The coherency measurement is an attribute called semblance Semblance is defined by the following quotient at s t z nq E where s t is the semblance trace and q t 7 is the j th sample on the on the input seismic trace In other words semblance is the square of the sum divided by n times the sum of the squares of the values on a seismic trace It should mentioned that there is more than one measurement of coheren
59. transform time to frequency of the suplane test pattern via suspecfx UNIX Quick Reference card pl From the University References UNIX Quick Reference card p2 naaa a Image of sonar su data no perc Only the largest amplitudes are visible Image of sonar su data with perc 99 Clipping the top 1 percentile of amplitudes brings up the lower amplitude amplitudes of the plot Image of sonar su data with perc 99 and legend 1 Comparison of the default hsv0 hsv2 and hsv7 colormaps Rendering these plots in grayscales emphasizes the location of the bright spot in the COlOMIAES G a s ace Boe Bnew ee ee ee ee ee ee a es te eae Image of sonar su data with perc 99 and legend 1 Image of sonar su data with median normalization and perc 99 Comparison of seismic su median normalized with the same data with no median normalization Amplitudes are clipped to 3 0 in each case Notice that there are features visible on the plot without median normalization that cannot be seen on the median normalized data Cartoon showing the simple shifting of time to depth The spatial coor dinates do not change in the transformation only the time scale t is stretched to the depth scale z Note that vertical relief looks greater in a depth section as compared with a time section a Test pattern b Test pattern corrected from time to depth c Test pattern corrected back from depth to time
60. water speed have been suppressed as have some of the multiples from the strong reflector near 2 sec and 2000 m s However this is not as clean as the radon transform filtered data We may repeat the process to eliminate other repetitions in the data such as those from pegleg multiples As with radon domain filtering we choose appropriate nmo velocities to flatten the arrivals we choose to remove You may want to try repeating the last several steps using the data fake water pegleg su 11 4 What else did predictive decon do to our data The fact that we are applying an inverse filter to our data means that in some sense we are making the output look more like a bunch of spikes or more like a bunch of Dirac delta functions Because we know that a spike contains all frequencies the term spectral whitening is applied to describe the effect of such filters in the frequency domain This bug feature may be observed in your data by comparing the amplitude spectra suspecfx lt fake water su suximage title data before spiking decon suspecfx lt pef faketwater su suximage title data after spiking decon On one hand it may seem that the increased frequency content is a good thing However can we really trust that those frequencies have been correctly added to the data These may be simply an artifact of the filter that causes more harm than good Some spectral whitening is desirable but most should probably be suppressed by
61. while vel 1t vmax do echo vel sunmo tnmo 0 0 vnmo vel lt data sustack gt gt movie vel expr vel dvel done suxmovie lt movie perc 98 n1 1500 n2 2142 loop 1 sleep 1000000 title fvel dvel exit 0 and ViewStack Movie bin sh movie stackmovie su fvel 1500 dvel 10 suxmovie lt movie perc 99 d1 004 n1 1500 n2 2142 sleep 10000 loop 1 title vnmo fvel exit 0 The idea of CVS is to perform constant velocity NMO on the data and stack it repeating for a large range of velocities Here this idea is implemented by sweeping through stacking velocities starting with vamo 1500 in increments of 10 m s The data are NMO corrected and then stacked Each stacked section becomes a frame in a movie In the title line of the movie the stacking velocity is expressed as vamo 1500 devel frame number It takes about 20 minutes on a modern multi core PC to generate the stackmovie su 192 12 5 Dip Moveout DMO Dip moveout is a partial migration that will convert NMO corrected prestack data to more closely approximate true zero offset data In the original formulation of dip moveout by Dave Hale the operation is applied to common offset data that have been NMO corrected The program sudmofk can be used to perform dip moveout on our NMO corrected data Hale s method was not amplitude preserving More modern applications of this type of data transformation that preserve amplitude have been de
62. 0 3650 dxcdp 12 5 gt stolt seis su exit 0 Here the pairs vmig are RMS velocities the velocities that come from velocity analysis stacking velocities for each time value in tmig These are only prelimi nary velocities taken from a single velocity analysis on the data If you look in the shell script PSmig you will see more PSmig bin sh sumigps lt seismic3 su tmig 0 0 1 0 2 5 3 45 4 36 5 1 5 45 5 95 vmig 1500 2000 3738 45 3338 3876 32 3678 11 5706 7 4156 17 dx 12 5 gt ps seis su exit 0 Here the pairs vmig are interval velocities the actual seismic wavespeeds for each time value in tmig The conversion of RMS to interval velocities is made through the application of the Dix equation In SU the program velconv is useful for a number types of velocities including rms to interval velocities The program suintvel is useful for converting a few stacking velocities to interval velocities For example you can run suintvel to convert the stacking velocities as a function of time used in the Stoltmig script into the interval velocities used in the PSmig script via suintvel t0 0 0 1 0 2 5 3 45 4 36 5 1 5 45 5 95 vs 1500 2000 3160 3210 3360 3408 3600 3650 106 The resulting velocities given by the value of v below h 0 1000 2803 84 1585 55 1763 72 1360 9 998 672 1039 04 v 1500 2000 3738 45 3338 3876 32 3678 11 5706 7 4156 17 The values of h are the depths corresponding to
63. 1 12 Setting up the working environment One of the most difficult and confusing aspects of working on Unix like systems is en countered right at the beginning This is the problem of setting up user s personal environment There are two sets of instructions given here One for the CSH family of shells and the other for the SH family 1 12 1 The CSH family Each of the shell types returned by SHELL has a different configuration file For the csh family tcsh csh the configuration files are cshre and login To configure the shell edit the file cshrc Also the path variable is lower case You will likely find a line beginning with set path with entries something like set path lib bin usr bin X11 usr local bin bin usr bin usr local bin usr sbin Suppose that the Seismic Unix package is installed in the directory usr local cwp on your system Then we would add one line above to set the CWPROOT environment variable And one line below to define the user s path setenv CWPROOT usr local cwp set path lib bin usr bin X11 usr local bin bin usr bin usr local bin usr sbin set path path CWPROOT bin Save the file and log out and log back in You will need to log out completely from the system not just from particular terminal windows When you log back in and pull up a terminal window typing echo CWPROOT will yield usr local cwp 21 and echo
64. 1200 2000 cdpt 1 8 2 2 trid 1 nhs 57 60 60 60 gelev 10 selev 6 scalel 1 scalco 1 SX 18212 28212 18212 28212 gx 15000 25000 15000 25000 counit 3 mute 48 ns 601 dt 4000 In each case where four numbers appear these are the minimum and maximum values in the header followed by the first and last values in the data panel You may use sukeyword to determine the meaning of any of the header field key words seen here via 56 sukeyword key where key is the specific keyword For example sukeyword tracl returns int tracl Trace sequence number within line numbers continue to increase if the same line continues across multiple SEG Y files The first field int tells us that this is defined as type integer in the header The short description is the SEG s definition for this field This can be a big deal Ofttimes users will want to define decimal values for the header fields Please note that the keyword names we use here are not an industry standard but are peculiar to SU These are an invention of Einar Kjartannson the author of the original suite of programs call SY that later became the basis of the SU package 4 3 2 Types of data formats In the world of scientific data there are three basic types of data formats These are acquisition internal and data exchange formats Acquisition formats An acquisition format is a data format that is natural to or conv
65. 14 Bally 14 5 TABEERTITAL nes HTT slopes 3 2 0 2 3 slopes 2 1 0 1 2 Figure 7 4 The results of a suit of Stolt migrations with different dip filters applied to limit the bandwith Stolt migration was applied to both the original unfiltered version of simple su and the frequency filtered version The frequency filtered version shows much less of the effect of spatial aliasing but also has lower resolution which is to say that the reflector is fatter This does not seem to be a problem on an image that has only one reflector but if there were a bunch closely spaced reflectors we could easily lose resolution of these reflectors with frequency filtering Dip filtering The third approach to spatial aliasing is filtering in the f domain Becuase we want to preserve as much of the frequency content of the data as possible the filter that we want to use should preserve the magnitude of the wavenumber vectors in the data but restrict the angles so that we suppress aliasing Such a filter is called a dip filter We may experiment with sudipfilt by trial and error sudipfilt dt 1 dx 1 lt simple su slopes 2 1 0 1 2 amps 0 1 1 1 0 sustolt cdpmin 1 cdpmax 80 dxcdp 40 vmig 2000 tmig 0 suxwigb xcur 3 title migration after dipfilter d2 40 amp sudipfilt dt 1 dx 1 lt simple su slopes 4 3 0 3 4 amps 0 1 1 1 0 sustolt cdpmin 1 cdpmax 80 dxcdp 40 vmig 2000 tmig 0 101 a 1000 2000 3000 b 1000 2000 3000
66. 21 124 126 129 130 136 10 3 a Suplane data b its Radon transform Note that a linear Radon trans form has isolated the three dipping lines as three points in the T p do main Note that the fact that these lines terminate sharply causes 4 tails on each point in the Radon domain 2 24 ote Bh ie eb 10 4 The suplane test pattern data with the steepest dipping arrival surgically removed in the Radon domain ooa a Be RE eke Baek eS 10 5 a Synthetic data similar to CDP 265 b Synthetic data plus simulated water bottom multiples c Synthetic data plus water bottom multiples plus select pegleg multiples 0a 0a a 000 2 ee eee nee 10 6 a Synthetic data similar to CDP 265 b Synthetic data plus simulated water bottom multiples c Synthetic data plus water bottom multiples plus select pegleg multiples 2 egg Ap aa Pe Bia eek Slave amp 10 7 a Synthetic data in the Radon domain b Synthetic data plus simulated water bottom multiples in the Radon domain c Synthetic data plus water bottom multiples plus select pegleg multiples in the Radon domain 145 10 8 CMP 265 NMO corrected with vnmo 1500 displayed in the Radon trans form T p domain Compare this figure with Figure 10 2 The repetition Indicates multiples sc ec Ye eee bee tee a eee 148 10 9 CDP 265 NMO corrected with the velocity function vamo 1500 1800 2300 tnmo 0 0 1 0 2 0 but with no stretch mute parameter applied NMO stretch artefacts appear in t
67. 2d performs the Kirchhoff migration drawing on the travel times created by rayt2d that are in the file tfile simple Indeed if a person had a better traveltime table generator then this program could use those traveltimes instead To apply the Kirchhoff migration to the dataset simple su we first type Rayt2d simple to generate the traveltime tables for sukdmig2d You will notice that a little window showing a movie will appear You may grab the lower corner of this window by clicking and dragging with your mouse to stretch the image to a larger size This movie shows the map of traveltimes to each point in the subsurface from the common source receiver position As this is only a constant background model we would expect that the curves of constant traveltime wavefronts are circles You should be able to detect this in the change of the grayscale If the curves do not appear to be perfect circles this is due the aspect ratio of the plot You may stretch the plot so that it has the correct aspect ratio To perform the the Kirchhoff migration we type Kdmig2d simple the resulting migrated data is the file kdmig simple su which you may view via suximage lt kdmig simple su amp 7 11 Spatial aliasing You may have noticed that the output from migrating the simple su test pattern no matter what migration method is used contains artifacts Because all of the various migration methods tried have more or less the same pattern of artifa
68. 4 isengard mines edu u 206424768 50614912 145324096 26 u isengard mines edu data cwpscratch 30963712 9732128 19658720 34 data cwpscratch isengard mines edu data 103212320 19418464 78550976 20 data 17 isengard mines edu scratch 396341312 199520 376008736 1 scratch the hardware devices on the far left column Those whose names begin with dev are hardware devices on the specific computer The items that begin with a machine name in this case isengard mines edu exist physically on another machine named isengard but are remotely mounted as to appear to be on this machine The second column from the left shows the total space on the device the third column shows the amount of space used while the fourth shows the amount available the fifth column shows the usage as a percentage of space used Finally the far right column shows the directory where these devices are mounted In Unix devices are mounted in such a way that they appear to be files or directories Under Unix like operating systems the user sees only a directory tree and not individual hardware devices If you try editing files in some of these other directories you will find that you likely do not have permission to read write or modify the contents of many those directories Unix is a multi user environment meaning that from an early day the notion of protecting users from each other and from themselves as well as protecting the operating s
69. 426 vnmo 1500 1654 88 1836 47 1971 5 2199 66 2683 91 Here some hand editing has been applied 12 2 1 Homework 9 Velocity analysis for stack Due 25 Oct 2012 before 9 00am Perform velocity analysis on the multiple suppressed data using Velan to obtain stacking velocities at a range of CDPs Alternatively you may use Velan radon if you like that script better These stacking velocities with corresponding cdp values will be in a file called nmovel par or radonnmovel par if you used Velan radon Apply NMO and Stack using sunmo par nmovel par lt radon gain jon 1 cdp su sustack gt stack nmo radon gain jon 1 su or sunmo par radonnmovel par lt radon gain jon 1 cdp su sustack gt stack nmo radon gain jon 1 su if you used Velan radon to do the picking Again the file names you use should be the names of your corresponding files Discuss the degree of improvement over the image quality given these better NMO velocities 12 2 2 Applying migration We then can apply any of the post stack migration algorithms that we have discussed earlier in the text Any of these corresponding tnmo vnmo pair sets may be copied from nmovel par a new file say stolt par and used as the par file as input for sustolt Select only one tnmo vnmo change tnmo and vnmo to tmig and vmig and you are ready to run sustolt For example sustolt par stoltmig par cdpmin 1 cdpmax 2142 dxcdp 12 5
70. 50 sx 3437 gx 1025 offset 2412 ep 109 cdp 50 sx 3412 gx 1050 offset 2362 ep 108 cdp 50 Sx 3387 gx 1075 offset 2312 ep 107 cdp 50 We notice a few characteristics of the data from the header fields First it seems that the cdp field is changing rapidly but eventually we see that we have more traces with the same cdp value What is happening is that the data do not have full fold on the beginning of the dataset We eventually have enough fold to have full coverage in CMP sx 6037 gx 3800 offset 2237 ep 213 cdp 265 sx 6012 gx 3825 offset 2187 ep 212 cdp 265 127 sx 5987 gx 3850 offset 2137 ep 211 cdp 265 Sx 5962 gx 3875 offset 2087 ep 210 cdp 265 Sx 5937 gx 3900 offset 2037 ep 209 cdp 265 sx 5912 gx 3925 offset 1987 ep 208 cdp 265 sx 5887 gx 3950 offset 1937 ep 207 cdp 265 9 6 5 Stacking Chart As we did in Section 9 4 2 we can use suchart to view the header fields graphically suchart lt seismic su keyl cdp key2 offset xgraph n 120120 linewidth 0 labell cdp label2 o0ffset marksize 2 mark 8 This is effectively a stacking chart in that it shows the available geophone positions for each CMP When data are stacked they are summed along lines of constant CMP on this chart If we zoom in on this plot then missing data becomes apparent as gaps in the plot of header values The missing shots are distributed over several CMPs and thus the effect of the missing data is minimized but not eliminated We do not have full fol
71. A Course in Geophysical Image Processing with Seismic Unix GPGN 461 561 Lab Fall 2012 Instructor John Stockwell Research Associate Center for Wave Phenomena copyright John W Stockwell Jr 2009 2010 2011 2012 all rights reserved License You may download this document for educational purposes and personal use only but not for republication November 1 2012 Contents 1 Seismic Processing Lab Preliminary issues 1 1 Motivation for thelabi oss sea oe ite ee a ae ae a RR 1 2 Unix and Unix like operating systems 1 2 1 Steep learning curve ce oh ae ae th ee a an oo ce Tees tS AsO BCI ie Ye a SE Se erty he ores A ad oe oe hy ER a Ge T4 Wines ae raset hen ble eee EES tas Ut ches Bam Oo deed wa ap Siete ge va 1 5 The working environment 2 4 4 9 4 4 4s ue Bek 4 Ane Ak pe ee he 1 6 Setting the working environment 2 040 E Choice Ol edi pOrite 4 4 Sek a ee Ald a een Seek SG Sh A Pe ee 1 8 The Unix directory structure a oe Gok ee Gober ees 1 9 Scratch and Data directories 000 0200022 eae 1 10 Shell environment variables and path 1 10 1 The path or PATH a a 1 10 2 The CWPROOT variable aoaaa Be bene Baek us 1 11 Shell configuration files aoa aa o 1 12 Setting up the working environment 1121 net H Aamil yas ie i tee BE ne Pa ee ee dee x Tee me family an Boe ase oe eA dee Be de ie 8 Ai ie Sg sen Se ee 1 13 Unix help m
72. CA 1 8 Auopadip U1 SJF Ist CF p s uondisaq pueuruiog snjejs JUBSWUOIIAUA AIQDLIDA JUAWUOAIAUA AAOUIAA A NIDA 01 ADA Aua Jay UOISS S JPU PUT pou ajowlad 01 USOT PUDU SDUD PUDUWUOI aaoway SDI D pUDUNUOI 2924A paomssvd asuvyy Zp SD pp Konoanp awvuay p Uoponp o1 f apf aaow p Uopanp aaoulay P K40jJa4IP MAU IDAJ p K40jJasIp 04 a8UuDYD uondrns aq U g upu paunu ausjesun 4 JUDU AUIS moog pu uou yawipu seieun Z upu j uimu sere pmssed cP Pp Aw P LUZI f su pP Tpu p ap yur pps pueuruio JO 1 U0D JUBWIUOJIAU UNIX Quick Reference card p2 30 Figure 2 4 Chapter 3 Lab Activity 2 viewing data Just as scratch paper is paper that you use temporarily without the plan of saving for the long term a scratch directory is temporary working space which is not backed up and which may be arbitrarily cleared by the system administrator Each computer in this lab has a directory called scratch that is provided as a temporary workspace for users It is in this location that you will be working with data Create your own scratch directory via mkdir scratch yourusername Here yourusername is the actual username that you are designated as on this system Please feel free to ask for help as you need it The scratch directory may reside physically on the computer where you are sitting or it may be remotely mounted In computer environements where the directory is locally on the a giv
73. E elas IH ea CT IN SPH OA MICE i function 0 1 2 velocity e stretch mute parameter d h 1e t 1000 A A Loh an a AJMAL bant eb adila Ll ete aed skat aiik Miik aebebbiela tasahitelaemian kil SLC a eet rth tini AHAH 1800 2300 tnmo 1500 No Stretch mute appl cdp 265 vnmo corrected with 265 NMO CDP vnmo 1500 1800 2300 tnmo 10 9 Figure 0O but with no 2 applied NMO stretch artefacts appear in the long offset 1 0 9 0 0 shallow portion of the section 153 you will note that there are two parameters smute 1 5 and Imute 25 which control the amount of stretch muting and the tapering of the mute We may run sumute with the stretch mute parameter turned off by choosing a large number for the value of smute sunmo vnmo 1500 1800 2300 tnmo 0 0 1 0 2 0 smute 8 lt gain jon 1 cdp 265 su The result is in Figure 10 9 We can clearly see the NMO stretch phenomenon The default values of smute and Imute work pretty well for most applications however if you see long period artifacts on your stacked section it is possible that you may need to adjust the values of the stretch mute Conversely we do have the possibility of losing data if the stretch mute is too agressive In either case the stretch mute may need to be adjusted 10 3 2 Muting specific arrivals The term muting simply means zeroing out parts of the data that we don
74. TEXT Entering Text Searching and Replacing a append after cursor lw search forward for w A or a append at end of line w search backward for w i insert before cursor wl n_ search forward for w and move down n lines Tor _i insert at beginning of line n repeat search forward o open line below cursor N repeat search backward o open line above cursor cm change text m is movement s oldiInew replace next occurence of old with new Cut Copy Paste Working w Buffers dm delete m is movement dd delete line D or d delete to end of line x delete char under cursor X delete char before cursor ym yank to buffer m is movement yy or Y yank to buffer current line p paste from buffer after cursor P paste from buffer before cursor bdd cut line into named buffer b a z bp paste from named buffer b s oldinew g replace all occurences on the line x ys old new g replace all ocurrences from line x to y s old new g replace all occurrences in file slold new gc same as above with confirmation Miscellaneous n gt m _ indent n lines m is movement n lt m un indent left n lines m is movement repeat last command U undo changes on current line u undo last command J join end of line with next line at lt cr gt xf insert text from external file f G show status Figure 1 1 A quick reference for the vi editor 16 You will see that your current working directory location which is your called your home dir
75. a clean up bin rm f radontmp exit 0 It takes several hours yes that s right hours perhaps 5 to 6 hours maybe longer to run this process on the full dataset We may view the results of the Radon transform multiple suppression by making an NMO correction to some relevant speed such as those used in assignment 5 and stacking to make a brute stack of the data for example sunmo vnmo 1500 2200 3000 tnmo 0 1 2 lt radon gain jon 1 cdp su key offset sustack suximage perc 99 amp If multiples are still prominent then we may need to perform the tau p filtering with better parameters or we may need to apply different filters to different ranges of CMPs 10 4 Homework Assignment 7 due Thursday 11 Oct 2012 before 9 00 AM Perform the following operations e Mute and perform gaining on your data then perform the analysis that you did in Homework 6 to a single CDP from each of these gathers to get tnmo and vnmo values for multiple suppression You will have a different set of tnmo and unmo values for each subset of the data e Break your data in to several parts via suwind lt gain jon 1 mute cdp su key cdp min 0 max 500 gt gain jon 1 cdp 0 500 su suwind lt gain jon 1 mute cdp su key cdp min 501 max 1000 gt gain jon 1 cdp 501 1000 su suwind lt gain jon 1 mute cdp su key cdp min 1001 max 1500 gt gain jon 1 cdp 1001 1500 su suwind lt gain jon 1 mute cdp su key cdp min 1501 max 2142 gt
76. aders on the original data that the spacing between midpoints which is to say the spacing between CMPs is 12 5m You may view the data via suximage lt seismic3 su perc 99 or suximage lt seismic3 su perc 99 verbose 1 The verbose 1 option will show you the actual clip values that perc 99 is giving you You may see what the actual values of bclip the numerical value of the color black and welip the numerical value of the color white bclip 0 569718 wclip 0 583124 You may then input clip values For example suximage lt seismic3 su clip 2 verbose 1 boosts the lower amplitudes Try different values clip to see what this does 8 1 Stolt and Phaseshift v t migrations The easiest migrations are the Stolt and Phaseshift migrations It is common to do a quick look at data by applying these types of migrations algorithms Because a velocity analysis has been applied to these data we have a collection of stacking velocities as a function of time which we will make use of as an estimate of migration velocity If you type 105 sustolt sumigps and look carefully at the parameters for these programs you will see that sustolt requires RMS velocities as a function of time whereas sumigps requires interval velocities If you look in the shell script Stoltmig you will see bin sh sustolt lt seismic3 su cdpmin 1 cdpmax 2142 tmig 0 0 1 0 2 5 3 45 4 36 5 1 5 45 5 95 vmig 1500 2000 3160 3210 3360 3408 360
77. amp view the output as wiggle traces This does exactly the same thing in terms of final output as the previous example with the exception that here two files have been created See Figure 2 2 27 2 2 1 Questions for discussion e What is the Fourier transform of a function e What is an amplitude spectrum e Why do the plots of the amplitude spectrum in Figure 2 2 appear as they do 2 3 Unix Quick Reference Cards The two figures Fig 2 3 and Fig 2 4 are a Quick Reference cards for some Unix commands 28 LLOYsqndsyin npabupipur amay ayyy 0 JOSMOIG MOA Jas QIM OPIN PHOM dy UO aps STY SS 0 OL 8661 Isn ny sI 31w peuondo ou 10 sox jqeuea pou Joyndw09 Joquunu OUIeUSTL JUOWUOIIAUS KIOPA x ssaid pue Aay Jonuoo umop poy X V Jajydwied siy ul pasn suoneinasqqy pied 30UAIIJAI SpueUIUOS xu SJOIAYAS ADOIONHOAL NOILVNAYOANYT ALISYAAINA 8086 Z L09 SOLIOJOOIIP WS S JOU 1dYJOSO POYUT OE SALIOJOAIIP ST SN ATUO SWOIYD UO PoT eISUT IEMS I SN ATUO UBD NOA WOO OUT POSSO AR NOA J IIAIMOH ESIDA DOTA pue eqo ur posso ae NOA UdYM UA SAY SUIOIYD NOK dU 10 ssadov ULI NOA yey SULIAU Wd SAs 1 POWYS SIYL Ojuiza eopl n N 1814 12 1S 20pI n Ny 1 AW xoyUa prnom ay Aroyaltp oJUTZA SIY 0 AIOPALIP OAS SY WOY O J AOU 0 payuem 20A UYor J afdurexa 104 SAY UM Syu MOQU S IJEU 10 SAJ AdOo SAY IAOU 0 SpUBLULHOD xu N ATeUTPIO I N ULI NOA PAS out p 330
78. ance between adjacent cdp bins m Optional Parameters noffmix 1 number of offsets to mix for unstacked data only tmig 0 0 times corresponding to rms velocities in vmig s vmig 1500 0 rms velocities corresponding to times in tmig m s Among the many parameters those that are required are cdpmin cdpmax dxcdp vmig and tmig Note the type of velocities vmig that this program requires are RMS or stacking velocities as a function of time To see the range of the cdps in the data type surange lt simple su 80 traces tracl 1 80 1 80 cdp 1 80 1 80 trid 1 ns 501 dt 4000 Thus cdpmin 1 and cdpmax 80 The data were made with a 40m spacing and with a velocity in the first medium of 2000m s Applying sustolt sustolt lt simple su cdpmin 1 cdpmax 80 dxcdp 40 vmig 2000 tmig 0 0 gt stolt simple su We may view the output of Stolt migration via suxwigb lt stolt simple su xcur 3 title Stolt migration of simple data amp The migrated data look similar in many ways to the graphical migration These data have been chosen to be too sparsely sampled spatially to be considered good This choice was made on purpose to accentuate the artifacts on the data These artifacts are known by the terms diffraction smiles migration impulse responses or endpoint contributions The fact that the data are sparse makes the migration operator see the individual arrivals as single spikes resulting in impulse responses o
79. annihilation method that makes assumptions about the structure of multiples based on an autoconvolution model of multiples Verschuur began with the observation that multiples could be made by the convolution of a seismic trace with itself suitably shifted and reasoned that it should be possible to use the data itself to model the multiples and then use adaptive subtraction to remove the multiples from the data The SRME method operates on a very simple model of multiples If we consider the auto convolution of data with itself then such resulting autoconvolved data are the same as multiples assuming simple single layer reflection For example 181 suconv lt fake su sufile fake su panel 1 suximage perc 99 title auto convolution o In theory we could use the earliest arrivals on a seismic reflection profile to build a model of the first multiples through autoconvolution and then adaptively subtract these out of the data The demultipled portion of the data would then be autoconvolved to generate a model of the second order bounces which in turn would be subtracted The process of modeling followed by adaptive subtraction can then be repeated until the data are completely cleaned of multiples as best as the algorithm could handle this Both 2D and 3D versions of SRME have been implemented Unfortunately we do not yet have an SRME code in the SU 11 10 Homework Assignment 8 Due Thursday 18 Oct 2012 before 9 00am This exercis
80. at the time sampling interval dt is expressed in microseconds Seismic fre quencies range from a few Hz to maybe 200 hz but likely are not up into the kilohertz range unless some special survey is being conducted Sonar frequencies likely range ten s of kilohertz to hundreds of kilohertz Radar operates in the megahertz range So it is common for the user to fake the units on the time sampling interval so as to fit the requirements of a seismic code 5 6 The sonar data The sonar su file is one of the profiles collected by Dr Henrique Tono of Duke Univer sity in a special laboratory setting 66 According to a personal communication by Dr Tono the geologic setting of the sonar data is thus The deposits and images were produced at the Saint Anthony Falls Lab of the University of Minnesota Here experimental stratigraphy is produced under precisely controlled conditions of subsidence base level and sediment supply By superimposing optical images of the sectioned deposits on seismic images we can directly observe the ability of seismic profiling to distinguish different geological features The experimental basin is 5 m by 5 m 25 m2 and 0 61 m deep Sediment and water were mixed in a funnel and fed into the basin at one corner This produced an approximately radially symmetrical fluvial system which aver aged 2 50 m from source to shoreline The edges of the basin were artificially roughened in order to direct
81. ata format Data may be in this format on tape media but today it is equally common for SEG Y data to be files stored on other media such as CD DVD or USB disk drive and memory stick devices These latter media are preferable for easy transport Our test dataset is stored as an SEG Y file called seismic segy in data cwpscratch Data5 as a data file in the SU format This file is big about 800 megabytes Make sure that you are working in an area on your system capable of storing several gigabytes You may copy this file to your working area via 111 cd scratch yourusername cp data cwpscratch Data5 seismic segy Temp5 9 1 1 Reading the data The data are in SEGY format and need to be read into the SU data data format This is done via cd Temp5 segyread tape seismic segy verbose 1 segyclean gt seismic su Reading seismic data particularly tapes is more of an art than a science If you ever request data from someone make sure that you get the data in a format that you can read Sometimes it is best to have them send you a small test dataset before committing to a larger set Seismic processing software whether commercial or open source has the property that there is an internal or working data format that usually differs greatly from the external or data exchange formats that data usually are transferred in In addition there are field data recording formats that differ still from the data exchange formats T
82. atch Data5 puts all of these together bin sh vnmo tnmo data pmula 0 pmulb 0 view nmo corrected data in the Radon domain sunmo vnmo vnmo tnmo tnmo lt data suradon offref 3237 interoff 262 pmin 2000 pmax 2000 dp 16 choose 0 igopt 2 depthref 1000 suximage perc 99 amp nmo gt radon gt inverse NMO for multiple suppression sunmo vnmo vnmo tnmo tnmo lt data suradon offref 3237 interoff 287 pmin 2000 pmax 2000 dp 8 choose 1 igopt 2 pmula pmula pmulb pmulb depthref 1000 sunmo vnmo vnmo tnmo tnmo invert 1 gt radon data 149 view semblance suvelan lt radon data dv 15 fv 1450 nv 200 suximage d2 15 2 1450 cmap hsv2 bclip 5 amp exit 0 10 2 1 Homework assignment 6 Due Thursday 4 Oct 2012 before 9 00am e Use suwind to capture a single CDP that is different from cdp265 that we have been studying in class suwind lt gain YOURPARAMETERS cdp su key cdp min NUMBER max NUMBER gt gain YOURPARAMETERS cdp NUMBER su Here for NUMBER is any any CDP number between 300 and 2000 View this file with suxwigb examine the headers with surange You are working with only one CDP here remember e Copy the shell script Test to your Temp5 directory cp data cwpscratch Data5 Test scratch yourusername Temp5 Modify this file to use your gained version of cdp NUMBER gain YOURPARAMTERS cdp NUMBER su file you created in in step 1 Run the shell scr
83. ater speed is about 5 s and we see stripes that are at about 51 s above and below the autocorrelation waveform spike Also we notice that there is an offset effect Thus we apply a moveout correction to flatten the data sunmo vnmo 1500 lt faketwatert tpegleg su supef maxlag 0673 suacor ntout 1024 suxwigb perc 90 Notice that the result is sensitive to the value of unmo It might be that making vnmo slightly bigger gives a slightly flatter collection of spikes sunmo vnmo 1800 lt faketwatertpegleg su supef maxlag 0673 suacor ntout 1024 suxwigb perc 90 The repetition time of the signal is the value that is needed to define the gap in the gapped decon In this case the gap is 51 seconds Our choice for minlag is a number that is slightly larger than the value of maxlag that we used for spiking the data Our maxlag for spiking decon will be the sum maxlag minlag gap Finally we finish by doing inverse NMO sunmo vnmo 1800 lt faketwater pegleg su supef maxlag 0673 supef minlag 068 maxlag 578 sunmo invert 1 vnmo 1800 suxwigb perc 99 xcur 2 We may view effect on the multiples by comparing semblance panels 175 suvelan lt faketwater pegleg su nv 150 dv 15 fv 1450 suximage d2 15 2 1450 cmap hsv2 bclip 5 title cdp 265 amp suvelan lt pef faket twatertpegleg su nv 150 dv 15 fv 1450 suximage d2 15 2 1450 cmap hsv2 bclip 3 title PEF amp The multiples with speeds near the
84. ation the performance as well as the accuracy at steeper dips of Gazdag migration can be improved The result is Dave Hale s sumigps sumigps SUMIGPS MIGration by Phase Shift with turning rays sumigps lt stdin gt stdout optional parms 84 Required Parameters None Optional Parameters dt from header dt or 004 time sampling interval dx from header d2 or 1 0 distance between successive cdps f fi1 0 0 0 5 dt 0 5 dt trapezoidal window of frequencies to migrate tmig 0 0 times corresponding to interval velocities in vmig vmig 1500 0 interval velocities corresponding to times in tmig vfile binary non ascii file containing velocities v t nxpad 0 number of cdps to pad with zeros before FFT ltaper 0 length of linear taper for left and right edges verbose 0 1 for diagnostic print Try an improved phase shift migration sumigps lt simple su dx 40 vmig 2000 tmig 0 0 gt ps simple su 7 4 Claerbout s 15 degree finite difference migration Another backpropagation approach was taken by Jon Claerbout in 1970 via finite difference solution of a one way approximate wave equation This is a reverse time finite difference migration but it is not the exact wave equation The 15 degree part refers to the angle of a cone about vertical within which the traveltime behavior of the migration is sufficiently similar to that of the standard wave equation as to yield an acceptable result Other approximations for
85. ay be thought of as the impulse response of the migration operation 6 3 Migration as a Diffraction stack Another approach to migration is immediately apparent If we apply Hagedoorn s method to the diffraction from a point scatterer then we observe that the scatterer is recon structed However tangent vectors are not defined with regard to a point scatter In stead it must be the ray vector from the source receiver position to the scatterer that is being reconstructed In other words the reflected ray vector is the distinguished vector associated with the imaging point For a reflector surface this is the perpendicularly reflected ray vector See Figure 6 7 Furthermore we might ask why is it necessary to draw Hagedoorn s circles at all Suppose that we were to sum over all possible diffraction hyperbolae Then the largest arrivals would exist only where a hyperbola we sum on hits a hyperbola in the data The sum would then be placed at a point at the apex of the hyperbola passing through our data This type of migration is referred to as a diffraction stack We sum or stack data but we do this over a diffraction curve Furthermore the output need not be a TT Figure 6 8 The light cone for a point scatterer at x z By classical geometry a vertical slice through the cone in x t the z 0 plane where we record our data is a hyperbola Time migrations collapse diffraction hyperbolae to their respect
86. be velocity analysis but unfortunately using sustolt we can only have a Upms t profile 199 14 2 Prestack Depth Migration Typically prestack depth migration algoritms assume that the data are not gained Prestack depth migration is a wave equation based process so the wave equation auto matically takes care of the effect of geometrical spreading The data also need to be in the form of shot gathers To undo geometrical spreading and resort the data into shot gathers susort sx offset lt myradon gain jon 1 cdp su sugain tpow 1 gt myradon shot su Similarly given a good background velocity profile unewvelxz bin For example sumigpreffd lt myradon shot su dx 12 5 dz 3 fmax 60 vfile unewvelxz bin gt premigffd radon su sumigpresp lt myradon shot su nxo 2142 nxshot 1001 nz 1500 verbose 1 dx 12 5 dz 3 fmax 60 vfile newvelxz bin gt premigffd radon gain jon 1 su sumigprepspi lt myradon shot su nxo 2142 nxshot 1001 nz 1500 verbose 1 dx 12 5 dz 2 fmax 60 vfile newvelxz bin gt premigffd radon gain jon 1 su These may take several days of runtime to complete 14 3 Concluding remarks Prestack depth migration is becoming increasingly possible on desktop hardware As a classroom exercise the cost is still prohibitive 200
87. brute stack as in Homework 5 of the full processed dataset and submit that instead Remember to show all commands and their parameters that perform actual processing steps Good for 5 extra points You might want to read about the at command in the next section before doing this assignment 10 4 1 The at command using the computer while you are asleep There is a famous adage that you cannot get rich unless you can find a way to make money while you are asleep In our case we don t efficiently use the computer unless we can run jobs when we are not physically at the machine such as when we are at home or asleep Such processing jobs are called batch jobs A batch job is a process that is submitted for later execution When computers were first invented all jobs were batch jobs 158 Batch jobs using the at command On Unix and Unix like systems the at command allows processing jobs to be run at a pre specified time by the user whether or not the user is logged in on the system The Unix man page for at man at shows the basic usage Suppose you have a shell script called Myshell located in some directory mydi rectory If you wanted to run this script in the middle of the night you could do the following cd mydirectory at lam tomorrow f Myshell An important point is that you need to have the f Today is 13 October 2011 for this discussion To see if your at job is on the list of jobs to be executed you wo
88. ctivity 13 Common offset gathers In the early days of seismic prospecting it was not unusual for surveys to be conducted of a source receiver geometry consisting of single shot geophone pairs collected at a common constant offset between source and receiver This natural because common offset gathers provide data that kind of look like an image of the subsurface the data image discussed in earlier chapters We view common offset gathers today as an important data sorting geometry From the Notes file we see that the minimum offset in the data is 262 m and the maximum offset is 3237 m We can save 3 common offset sections via 122 suwind lt seismic su key offset min 262 max 262 sugain jon 1 gt gain jon 1 offset 262 su suwind lt seismic su key offset min 1012 max 1012 sugain jon 1 gt gain jon 1 offset 1012 su suwind lt seismic su key offset min 3237 max 3237 sugain jon 1 gt gain jon 1 offset 3237 su Note that we have used jon 1 for convenience You should use your own values of tpow gpow and qclip View and compare these respective near intermediate and far offset sections Note the presence of multiples and the time resolution on each section as well as the time of the first arrival representing the water bottom Indeed some operations such as prestack Stolt migration sustolt and FK and TX Dip moveout sudmofk and sudmotx require that the input data be resorted into common offset
89. cts we are led to suspect that this is caused by the simple su dataset rather than the migration routines themselves We can study the problem by migrating a different version of the test pattern called interp su cd scratch yourusername Temp2 cp data cwpscratch Data2 interp su suxwigb lt interp su xcur 3 title interp data surange lt interp su surange lt simple su You will notice that interp su appears to show the same data as the original simple su data but surange shows that there are 159 traces instead of the 80 traces that were in simple su The interp su data were made by using the program suinterp interpolate the traces in simple su The interpolation was done via the command suinterp lt simple su sushw key tracl cdp a 1 1 b 1 1 gt interp su 94 The sushw program fixes the header values so that the trace numbers are accurately represented If we run sustolt to migrate the interp su data and compare this with performing Stolt migration on the original simple su data sustolt lt interp su cdpmin 1 cdpmax 159 dxcdp 20 vmig 2000 tmig 0 suxwigb xcur 3 title interpolated sustolt lt simple su cdpmin 1 cdpmax 80 dxcdp 40 vmig 2000 tmig 0 suxwigb xcur 3 title simple data then we see that the interpolated data yield a much better image Please note that cdpmax 159 and because the data are interpolated the spacing between traces is dxcdp 20 which is half of the value used for the previous
90. cy that geophysicists employ but semblance is the most commonly used measure The program that does this is suvelan suvelan nv 150 fv 1450 dv 15 lt gain jon 1 cdp 265 su suximage d2 15 2 1450 verbose 1 cmap hsv2 legend 1 bclip 5 amp Running this command on gain jon 1 cdp 265 su we can see evidence of multiples which need to be suppressed before we can proceed further We see that the multiples have slow moveouts One set of multiples tends to approximate the water speed Another set shadows strong reflectors in the subsurface 135 r 500 2000 2500 3000 3500 0 4 0 3 0 2 0 1 Figure 10 1 Semblance plot of CDP 265 The white dashed line indicates a possible location for the NMO velocity curve Water bottom multiples are seen on the left side of the plot Multiples of strong reflectors shadow the brightest arrivals on the NMO velocity curve 136 offset meters 2000 1000 3000 Kd Ad bad da i La b TS TT h da Brah ak Hi abet da Lakisi bl AE Whi inti dard Meee a H Py OMe er ae ee Ee riser es ae KAANE rel Ae ae a nr ee et th ei miari asia bare th tht Th Wine i dar eather iw on a Aaa vette Si ee ate pe ee Tiasa iala HHS tee NY Ter Ot eek ee AAEN WHA We Pah IYA R end seas Rie re ott eel ELLEF RI er ites ES EM oat th vnmo 1500 CMP 265 w 1500 Arrivals that we want to keep is horizontal or curves down Figure 10 2 CMP 265 NMO corrected with vn
91. d 25 2 gt WV Des Or dala f rmats scis te of aoe SP ok Se amp et pe ee S 4 4 Concluding Remarks a cays elie amp ae Se deh ead Mule Se Be Al gets Sse Lab Activity 4 Migration Imaging as depth conversion 5 1 Imaging as the solution to an inverse problem 5 2 Inverse scattering imaging as time to depth conversion 5 3 Time to depth with suttoz depth to time with suztot 5 4 Time to depth conversion of a test pattern 0 2 4 5 4 1 How time depth and depth time conversion works 5 5 Sonar and Radar bad header values and incomplete information 50 ANE Sonar Gala lt 2 maes oe 8 a ie Ara oo be SO hee She ew ee ok a 5 7 Homework Problem 2 Time to depth conversion of the sonar su and the radar su data Due Thursday 6 Sept 2012 0 22 5 8 Concluding Remarks 2 fsa gs a Be aoe esa pe eee BO te a amp ee ees Zero offset aka poststack migration 6 1 Migration as reverse time propagation 0 4 6 2 Lab Activity 5 Hagedoorn s graphical migration 6 3 Migration as a Diffraction stack ac2d ii ok eh ee ete Se SI Gee Fe wx 6 4 Migration as a mathematical mapping 4 kon ale gs a se a a ee a 6 5 Concluding Remarks 4 0 4 oan edd e alee dae led oa le ee eos Lab Activity 6 Several types of migration 7 1 Different types of velocity 2 a 540 4 5 tal 3 odd eS Re aS tah 7 1 1 Velocity conversion v t to v z
92. d on the 120 CMPs on the ends of the data but you can see that on low CDP number end of the plot it is mainly far offset data that are stacked whereas on the high CMP number side of the data it is near offset data that are stacked This accounts for the strange appearance of the edges of the data that we see in stacked sections 9 6 6 Capturing a Single CMP gather Around cdp 265 we are near the ep 200 portion of the data We can capture this CMP gather with suwind suwind lt gain jon 1 cdp su key cdp count 120 min 265 max 265 gt gain jon 1 cdp 265 su which may be viewed as wiggle traces suxwigb lt gain jon 1 cdp 265 su For a better view we may plot the wiggle traces in true offset reading the offset from the header field suxwigb lt gain jon 1 cdp 265 su key offset amp revealing that there are missing traces due to the missing shots in the data 128 offset meters 2000 1000 3000 tal i N Heal aa ri i AF k y a LOTTE Er K or Veith MEL ay h v i GETE AAO Ta ads iB hin to i Nias Pa hit brhth HF MEE HEFEH eh N Vilisilr Cha ct 4 Me N ae dtd bt A LATORI a CDP 265 CMP 265 of the gained data 6 9 igure F 129 a midpoint m x104 b midpoint m x104 3 1 2 3 4 5 CV Stack vnmo 1500 c midpoint m x104 d midpoint m x104 1 2 3 4 5 1 2 3 Raw Stack no NMO time s Sialk vnmo 2300 Brute Stack vnm
93. data The cascade of these processes would be a migration followed by remodeling to the desired zero offset traces In fact you could remodel to any type of data But if you could do this why bother With all of those integrals it would be horribly expensive in computer time The answer is that a number of the integrals in the cascade of migration remodeling can be done approximately using asymptotic methods so that the resulting algorithm looks like a migration algorithm but migrates the data not to the earth model but to approximate zero offset data These data could then be stacked and then migrated with a conventional migration program One motivation for doing this is to create a more advanced type of velocity analysis The usual NUO DMO STACK procedure is only approximate for real data The TZO methods work better though they are more expensive In fact there are a number of forms of migration velocity analysis that make use of the ideas stated here The other motivation was to apply DMO as a substitute for full 3D depth migration back when computer speed and storage was more limited 194 Chapter 13 Velocity models horizon picking and muting One seldom uses velocity models derived directly from velocity analysis What is more common is to use a depth migrated seismic image combined with an estimate of interval velocities to construct a background wavespeed profile that is used for either poststack or prestack de
94. data with the respective dip filters applied in the Stolt migrations of Figure 7 4 The first 1000 traces in the data ied sit a fo Bo Sent od he gh asked he Sede kd a Shot 200 as wiggle traces b as an image plot Gaining tests a no gain applied b tpow 1 c tpow 2 d jon 1 Note that in the text we often use jon 1 because it is convenient not because it is optimal It is up to you to find better values of the gaining parameters Once you have found those you should continue using those Common Offset Sections a offset 262 meters b offset 1012 meters c offset 3237 meters Gaining is done via sugain jon 1 A stacking chart is merely a plot of the header CDP field versus the Affect field Note white stripes indicating missing shots CMP 265 of the gainedsdatas vinta ct s 3 e k So ce He A ee oe eS CV stacks and a Brute stack a Raw stack no NMO correction b vnmo 1500 c vnmo 2300 d vnmo 1500 1800 2300 tnmo 0 0 1 0 3 0 Semblance plot of CDP 265 The white dashed line indicates a possible location for the NMO velocity curve Water bottom multiples are seen on the left side of the plot Multiples of strong reflectors shadow the brightest arrivals on the NMO velocity curve 2 000 CMP 265 NMO corrected with vnmo 1500 Arrivals that we want to keep curve up wheres multiple energy is horizontal or curves down 96 98 100 101 102 115 117 1
95. deviate from this assumption For air guns there is a reverberation known as a bubble pulse which is caused by a reverberation of the air bubble generated by the air gun The second problem with marine data is known as ghosting Ghosting is caused because there is are reflections off of the water surface at both the source and the receiver that tend to turn the waveform into something more like a doublet or triplet Ghosting is evidenced by a notch in the spectrum of the seismic data These deviations from perfect minimum phase are handled by deconvolution Indeed we may also use predictive deconvolution to suppress multiples 10 3 7 Further processing We may consider processing the full dataset with the shell script Radon final bin sh data gain jon 1 cdp su radondata radon data these nmo times and velocities are obtained by doing multiple Radon test runs vnmo 1500 tnmo 0 0 interoff 262 of fref 3237 depthref 1000 pmin 2000 pmax 2000 dp 8 igopt 2 you might consider making these values 10 or something we rely on the choice of tnmo and vnmo to separate data and multiples pmula 0 pmulb 0 156 filter data in radon domain choose 1 data multiples sunmo vnmo vnmo tnmo tnmo lt data suradon offref offref depthref depthref pmula pmula pmulb pmulb interoff interoff pmin pmin pmax pmax dp dp choose choose igopt igopt sunmo vnmo vnmo tnmo tnmo invert 1 gt radondat
96. difference migrations sumigffd lt infile gt outfile vfile optional parameters Required Parameters nz number of depth sapmles dz depth sampling interval vfile name of file containing velocities Optional Parameters dt from header dt or 004 time sampling interval dx from header d2 or 1 0 midpoint sampling interval ft 0 0 first time sample fz 0 0 first depth sample Try 86 cp data cwpscratch Data2 vel fdmig simple scratch yourusername Temp2 cd scratch yourusername Temp2 sumigffd lt simple su dx 40 dz 12 nz 150 vfile vel fdmig simple gt ffd simple su 7 6 Stoffa s split step migration Another algorithm known as the split step algorithm developed by P Stoffa et al in 1990 is an extension of this idea with sumigsplit SUMIGSPLIT Split step depth migration for zero offset data sumigsplit lt infile gt outfile vfile optional parameters Required Parameters nz number of depth sapmles dz depth sampling interval vfile name of file containing velocities Optional Parameters dt from header dt or 004 time sampling interval dx from header d2 or 1 0 midpoint sampling interval ft 0 0 first time sample fz first depth sample Try cp data cwpscratch Data2 vel fdmig simple scratch yourusername Temp2 cd scratch yourusername Temp2 sumigsplit lt simple su dx 40 dz 12 nz 150 vfile vel fdmig simple gt split simple su T T Gazdag s Phase shift Plus Interpolation mi
97. domain These are topics discussed in later chapters In the industry a method called surface related multiple elimination SRME is very popular This method models multiples as the autoconvolution of pri mary reflections permitting the multiples to be modeled and subtracted out leaving the data Clearly we are not finished and in fact we have not really gotten very far yet 9 9 Concluding Remarks After Harry Mayne patented the CDP stacking method in 1950 as part of his work at the Petty Geophysical Engineering Company oil companies were required to pay a licensing fee to Petty to use the technique commercially The technique of sorting NMO correcting and stacking the data was done with the data in analog form This required highly specialized and multi milllion dollar magnetic drums and magnetic tape devices Such operations as NMO and STACK were performed using this unwieldy and tempermental equipment The ease and simplicity of applying these operations on digital data as we do in lab assignments was years away These operations were costly and technically 133 difficult in the pre digital era yet even with only 6 fold stacking the improvements in data quality were worth it Rather than pay Petty the licensing fee some companies instead did the common depth point sorting part but did not do the stack Instead of stacking the data some companies developed high density plotters such that the traces were effectively plotted one on
98. done to the data and in fact some of these operations should be done before gaining 132 Muting Some seismic arrivals should be removed before attempting to gain the data Muting out direct arrival energy is one such item Muting means to zero out the data in specific ranges of space and time The SU program sumute allows the muting operation to be performed What do we mute We mute direct arrivals and the place where direct arrivals interact with reflections and refracted arrivals at the earliest times of the data There is also something called a stretch mute which is the removal of a low frequency artifact of the NMO correction that is viewed on NMO gathers Wavelet shaping One of the first things that we do to data is to correct for the shape of the wavelet This is done by deconvolution There are a number of tools that we may use to improve the waveform that involve methods with a variety of assumptions We discuss some of these in depth in Chapter 11 Multiple suppression As our brute stacks show particularly stacks with vamo 1500 that accentuates energy traveling at or near the water speed our data are dominated by water bottom and peg leg multiples We need to do gaining to make the amplitudes of the multiples more uniform in order to remove them but then we must re gain the data Our methods of multiple suppression include predictive deconvolution and filtering in the tau p also known as the Radon or slant stack
99. dowed interfaced version of emacs called xemacs that is similar to the first two editors These are all easy to learn and to use Not all editors are the best to use The user may find that invisible characters are introduced by some editors and that there may be issues regarding how wrapped lines are handled that may cause problems for some applications More advanced issues such as those which might be of interest to a Unix system administrator narrow the field back to vi The choice of editor is often a highly personal one depending on what the user is familiar with or is trying to accomplish Any of the above mentioned editors or similar third party editors likely are sufficient for the purposes of this course 1 8 The Unix directory structure As with other computing systems data and programs are contained in files and files are contained in folders In Unix and all Unix like environments folders are called directories The structure of directories in Unix is that of an upside down tree with its root at the top and its branches subdirectories and the files they contain extending downward The root directory is called pronounced slash While there exist graphical browsers on most Unix like operating systems it is more efficient for users working on the commandline of a terminal windows to use a few simple commands to view and navigate the contents of the directory structure These comma
100. e Se E a UA Ae ae the S ae hove of 76 TT 7 1 ie 7 3 7 4 7 5 9 1 9 2 9 3 9 4 9 5 9 6 9 7 10 1 10 2 a The simple su data b The same data trace interpolated the interp su data You can recognize spatial aliasing in a by noticing that the peak of the waveform on a given trace does not line up with the main lobe of the neighboring traces The data in b are the same data as in a but with twice as many traces covering the same spatial range Each peak aligns with part of the main lobe of the waveform on the neighboring trace so there is no spatial aliasing 2 20 24 20254 26 44574525 a Simple data in the f amp domain b Interpolated simple data in the f amp domain c Simple data represented in the kz kz domain d In terpolated simple data in the k k domain The simple su data are truncated in the frequency domain with the aliased portions folded over to lower wavenumbers The interpolated data are not folded a simple su data unfiltered b simple su data filtered with a 5 10 20 25 Hz trapezoidal filter c Stolt migration of unfiltered data d Stolt migra tion of filtered data e interpolated data f Stolt migration of interpolated data Clearly the most satisfying result is obtained by migrating the in terpolatcdditas a ule eae Bee ee at we ae ee eee oO The results of a suit of Stolt migrations with different dip filters applied The k1 k2 domain plots of the simple su
101. e data Consider for example data made by running suplane in the following suplane ntr 120 nt 256 dt 004 sushw key offset a 0 b 10 sufilter f 0 5 50 60 gt suplanedata su suximage lt suplanedata su title suplane data key offset labell time s label2 offset meters where we have 3 intersecting linear arrivals We would like to separate the data in such a way that will allow us to remove one of the dipping arrivals We could use dip or slope filtering in the f k domain to do this In this case filtering in the f k domain might be the best thing owing to the fact that these are straight lines However there is another method called the Radon or T p transform that will do a better job of separating arrivals particularly if the arrivals are curves with different curvature such as we see with arrivals having differing moveout The Radon transform operates by summing the data over curves that are drawn from time at an intercept offset interoff which is usually usually either 0 or the smallest offset in the data The curves are referenced an offset offref that is usually taken as the maximum offset in the data The data are summed over a fan of these curves taken with differing initial slope and the result plotted as a function of the reference time the T and slope the p of each curve For example transforming the suplanedata su made above into the Radon domain is done with suradon 138 a offset meters p s
102. e drawn We must not expect more out of our data than we have a right to expect but we must also not give up too easily 160 Chapter 11 Spectral methods and advanced gaining methods for seismic data processing There is a class of methods that are best called spectral methods because they modify the amplitude and or phase spectrum of the data There are several reasons for performing such operations The first is to sharpen the waveform to more clearly define reflection arrivals The second is to perform multiple suppression by identifying and suppressing repetitions in the data Some of these techniques are filtering operations but others are best thought of as a deconvolutional process that is to say a process by which data are inverted in some sense There are many such methods but we will discuss only a small subset of these so you get the idea of the kinds of things that processors apply Such processes are generally performed before velocity analysis as to suppress multi ples and to make the waveform sharper as to improve the resolution on the semblance plots that are used for velocity analysis 11 1 Common assumptions of spectral method processing Seismic data result from the introduction of seismic energy into the subsurface followed by the subsequent recording of reflections either on the surface of the earth or in a well bore Such data have a natural frequency band and a natural phase spectrum We may seek to change
103. e if the gaining is correct 9 5 3 Statistical gaining Another commonly used method is automatic gain control AGC The notion of having some automatic or programmed amplification factor dates from the days of analog record ing Instruments have a maximum range of amplitudes that they can record known as the dynamic range of the instrument It was common that the dynamic range of seismic recording instruments could not accomodate the full range of seismic amplitudes Adjust the instrument so that it could record all of the early larger amplitudes without clipping and the smaller and later amplitudes would be lost Set the instrument for the smaller amplitudes and the recording system would be overwhelmed by the larger amplitudes at the earlier times in the data The solution was to use a different gains in differing time windows in the data 120 a offset meters x104 b offset meters x104 1 130 1 132 1 134 1 136 1 138 1 140 1 130 1 132 1 134 1 136 1 138 1 140
104. e in GPR or near surface engineering geo physics applications may expect data to be written in the SEG 2 or SEG B formats Data excahnage formats For data to be shared between companies or other users yet a third class of data format is required Such formats are called data exchange formats The most popular is SEG Y though it is possible that data in SEG D SEG B SEG 2 or other format Any format that is relatively stable may effectively become a data exchange format whether or not the originators of that format had this in mind The SU data format is treated as a data exchange format by some software developers 4 4 Concluding Remarks Every data processing package has help features and internal documentation None of these are usually perfect and all are usually aimed at people who already understand the package Look for the help features and demos of a package When receiving a dataset the most important questions that a scientist can ask about a dataset that he or she receives are What is the format of the data Are the data uniformly sampled For seismic data Have the headers been set and What are ranges of the the header values Do the header values make sense Note also that data coming in from the field frequently requires that the headers be set in the data Transferring header information from seismic observers logs into seismic trace headers is called setting geometry Setting geome
105. e is similar to problem 7 e Start with the ungained and unmuted data e Mute and gain dataset e Apply one of the spectral methods to sharpen the waveform of the data You may wish to attempt to use predictive decon to suppress the near offset multiples that are not suppresed by filtering in the Radon transform domain e Use suwind to break the full dataset into blocks that are about 500 CDPs in size The number of blocks is up to you For example the last block was more than 500 CDPs so you could split that one up e Perform the Radon domain multiple suppression on each of these smaller blocks to do a better job of multiple suppression It is more important to preserve the real reflections than it is to try to suppress all multiples at the expense of data You may have done all of this already in the previous assignment Redo parts only if feel that you need to improve the result e Concatenate the blocks together to form the full multiple suppressed version of the dataset Call this dataset radon gain yourgainparameters cdp su e Repeat exercise 5 on this multiple suppressed dataset Save your NMO velocities in a file called say radonnmovel par For example if you used the CDPs at 250 750 1250 1750 to get your NMO velocities then the contents of the file would look something like 182 cdp 250 750 1250 1750 ENMO O0 55 ag ape ages bay Oe vnmo 1500 32050 tnmo 0 vnmo 1500 tCNMO O ose ey ede
106. e number of samples These are all things which we will need for further processing 9 2 1 Setting geometry One of the most time consuming and difficult and yet one of the most important steps in reading seismic data sets occurs in this step of the process This is called setting geometry The process is one of converting field observation parameters recorded in the field observers logs into trace header values The process itself is often time consuming if everything is correct in the logs but typically there are errors in observers logs that complicate this process It can take as long as a month to set the geometry in a 3D dataset In SU the tools for setting geometry are suaddhead sushw suchw sudumptrace suedit suxedit suutm A A A RA BA BRA HA We may make use of sushw and suchw later in the notes For the most part we will assume that our dataset has had geometry set properly 113 9 3 Getting to know our data Viewing the data If we know that our data have the trace headers set correctly the next part of working with the data is to view subsections of the data to see if there are missing traces zero traces and bad traces We are interested in the quality and reproduceability of the data across the section We are also interested evaluating whether there is noise that may need to be suppressed 9 3 1 Windowing Seismic Data It is always a good idea to look at some small part of the data to see if you have data
107. e r c In the real earth the wavespeed varies with position so the wavefront surface will not be a simple spherical shell it will be a complicated function There may be focusings and defocussings due to the velocity variations that cause the wave amplitude to decay by a function that is more complicated than a simple 1 r In reality the divergence correction problem is a problem of modeling wave amplitudes as a function of the velocity model Anelastic attenuation The second cause of wave amplitude reduction is due to anelastic attenuation Rock may be thought of as a kind of spring that undergoes distortion in a cyclical fashion as a wave travels through it Owing to frictional as well as due to more complicated effects involving the motion of fluids in rock pore spaces some of the seismic wave energy is converted to heat This failure of the material to behave exactly elastically is is called 1The reader may have expected an inverse square law from experiences in undergraduate physics classes rather than a 1 r law Energy density does diminish according to an inverse square law but because seismic wave energy is proportional to the square of seismic amplitudes the 1 r amplitude loss is consistent with an inverse square law of energy spreading 118 anelastic attenuation There is an additional loss of energy due to scattering which is called scattering attenuation The effect of anelastic and scattering attenuatio
108. e reflector by defining the tangent vectors of the reflector What then are the circles we have drawn The answer can by found by looking at Figure 6 5 For our 2D constant wavespeed example all solutions of the wave equation which is to say all wavefronts can be found by passing a horizontal plane through the cone in Figure 6 5 Both physical causal solutions the positive t cone and the nonphysical anti causal solutions the negative t cone are depicted We use the causal cone for modeling and the anti causal or reverse time cone for migration To see what a given circle means in Hagadoorn s method we may look at the reverse time cone in Figure 6 6 We may think of the curve on the t 0 plane as the locus of all possible positions from which the reflection originated or we may think of this as the wavefront of the backward propagated wave If we were to apply the Hagedoorn method on the computer we might consider creating for each seismic trace a panel of seismic traces replicating our original seismic arrivals but on a semicircular pattern Spraying out our seismic data for each trace along the respective Hagedoorn circle would yield one new panel of traces for each seismic trace Our 80 traces would then become 80 panels of sprayed traces We would then sum the corresponding traces on each panel Constructive interference would tend to enhance the region near the reflector and destructive interference would tend to eliminate e
109. e shell The user has access to an application called terminal in the graphical user environ ment that when launched usually by double clicking invokes a window called a terminal window The word terminal harks back to an earlier day when a physical device called a terminal consisting of a screen and keyboard but no mouse constituted the users interface to the computer It is at the prompt on the terminal window that the user has access to a commandline where Unix commands are typed Most commands on Unix like systems are not built in commands in the shell but are actually programs that are run under the users working shell environment The shell commandline prompt is asking the user to input the name of an executable program 1 5 The working environment In the Unix world all filenames program names shells and directory names as well as passwords are case sensitive in their input so please be careful in running the examples that follow If the user types cd lt change directory with no argument takes the user to his her home don t type the dollar sign directory In these notes the symbol will represent the commandline prompt The user does not type this Because there are a large variety of possible prompt characters or strings of 13 characters that people use for the propmt we show here only the dollar sign as a generic commmandline prompt On your system it might be a a gt or so
110. echanism Unix man pages 0 Lab Activity 1 Getting started with Unix and SU 2 1 Pipe redirect in lt redirect out gt and run in background amp 2 2 Stringing commands together 0 lt l 4 4 ute wea Boe ko AS aes 2 2 1 Questions for discussion 4 5508 oa ibd dS Sw ode Bese a ee aes 2 3 Unix Quick Reference Cards hue 6 3 kia oe Ae eee Beas ae hed Lab Activity 2 viewing data 3 0 1 Data image examples 2 8 eof alee se ee ea eee 3 1 Viewing an SU data file Wiggle traces and Image plots 3 1 1 Wiggle traces a At Ee ts Bh Sees oF a Sul Imag plots en Laas eet a hie ee A tee EL were de oh as ee ae 3 2 Greyscale lt 2 i so a tras oer ee Skee e eee 4 CRONE amp et pe ire amp 3 3 Legend making grayscale values scientifically meaningful 3 4 Normalization Median balancing 2 004 11 11 12 12 12 13 13 14 14 15 18 19 20 20 20 21 21 22 22 24 26 27 28 28 3 5 Homework problem 1 Due Thursday 30 August 2012 3 6 Concluding Remarks Yu oY Biens do Bea Ske ken ay oo be dtees atin be giee 3 6 1 What do the numbers mean 0004 Help features in Seismic Unix 4T CPG SC AGG et 82 Fas oh ee ete ede i OE OO ght ASOD ge Ate SY Se A R 4 2 Finding the names of programs with suname 4 3 Lab Activity 3 Exploring the trace header structure 4 3 1 What are the trace header fieldsssukeywor
111. ectory You should see something like pwd home yourusername where yourusername is your username on the system Other users likely have their home directories in home or something similar depending on how your system administrator has set things up The command Is will show you the contents of your home directory which may consist of files or other subdirectories The codes for Seismic Unix are installed in some system directory path We will assume that all of the CWP SU Seismic Unix codes are located in usr local cwp This denotes a directory cwp which is the sub directory of a directory called local which is in turn a directory of the directory usr that itself is a sub directory of slash It is worth wile for the user to spend some time learning the layout of his or her directories There is a command called df which shows the hardware devices that constitute the available storage on the users machine A typical output from typing df Filesystem 1K blocks Used Available Use Mounted on dev sdal 295101892 45219392 234892216 17 none 4070496 268 4070228 1 dev none 4075032 4988 4070044 1 dev shm none 4075032 124 4074908 1 var run none 4075032 O 4075032 0 var lock none 4075032 O 4075032 0 lib init rw isengard mines edu usr local cwp 20314752 6037440 13228736 32 usr local cwp isengard mines edu usr local sedpak54 20314752 6037440 13228736 32 usr local sedpak5
112. ed to combat spatial aliasing provided that all frequencies are not spatially aliased in Fig 7 3 there is a comparison with such a frequency filtered case The simple data were filtered via the program sufilter sufilter lt simple su 0 5 20 25 gt simple_filtered su 99 a trace number 40 time s simple su unfiltered trace number c 20 ll time s e wees A e i Spe EI Ue v simple su filtered trace number 100 150 simple su interpolated A time s trace number 20 40 60 80 a
113. elzx bin and unewvelxz bin which may then be used for migration time 13 2 Migration velocity tests Likely you will not get the correct migration velocities on the first iteration of the velocity modeling process The shell script Gbmig bin sh Gaussian Beam set x use a v z x velocity profile here vfile unewvelzx bin vary the values of dz and nz as a test before running the full model sumiggbzo lt stack dmo nmo myradon gain jon 1 su dx 12 5 dz 3 nz 1500 verbose 1 vfile vfile gt gb seismic su exit 0 applies Gaussian beam migration to the stacked data using the velocity file unewvelzx bin This migration result may be viewed with suximage via suximage lt gb seismic su perc 99 legend 1 d1 3 d2 12 5 amp and the errors in migration velocity may be seen as either frowns indicating under migration or smiles indicating over migration of the data The user may hand edit the file unif2 par increasing or decreasing the velocities in response to the respective smiles or frowns The user then runs Unif2 again and re runs the migration This process is repeated until the image has neither smiles nor frowns 197 13 2 1 Homework 11 Build a velocity model and perform Gaussian Beam Migration Due 15 Nov 2012 Use the methods outlined in this chapter to build a velocity model and perform Gaussian beam migration using that model e Use the RMStoINT and Suttoz shell scripts to stretch the Stolt mig
114. en computer you will have to keep working on the same system If you change computers you will have to transfer the items from your personal scratch area to that new machine In labs where the directory is remotely mounted you may work on any machine that has the directory mounted Remember scratch directories are not backed up If you want to save materials permanently it is a good idea to make use of a USB storage device 3 0 1 Data image examples 9 06 Three small datasets are provided These are labeled sonar su radar su and seis mic su and are located in the directory data cwpscratch Datal We will pretend that these data examples are data images which is to say these are examples that require no further processing Do the following cd scratch yourusername this takes you to scratch yourusername 31 This represents the prompt at the beginning of the commandline Do not type the when entering commands mkdir Temp1 this creates the directory Temp1 cd Tempi change working directory to Temp1 cp data cwpscratch Data1 sonar su cp data cwpscratch Data1 radar su cp data cwpscratch Data1l seismic su This is a literal dot which means the current directory 1s should show the file sonar su AAA HAHA For the rest of this document when you are directed to make Temp directories it will be assumed that you are putting these in your personal scratch directory
115. en to fit on the page When you type this the pipe follows immediately after the seismic su There are other possibilities We may consider simply normalizing the data by the maximum or minimum value or by some other constant Furthermore we have the question of whether the process be applied trace by trace or over the whole panel of data 3 5 Homework problem 1 Due Thursday 30 August 2012 Repeat display gaining experiments of the previous section with radar su and seis mic su to see what median balancing and setting perc does to these data 40 200 400 0 05 0 25 Figure 3 6 Image of sonar su data with median normalization and perc 99 41 68357 9 0 30397 999 0 68000 median no median Figure 3 7 Comparison of seismic su median normalized with the same data with no median normalization Amplitudes are clipped to 3 0 in each case Notice that there are features visible on the plot without median normalization that cannot be seen on the median normalized data 42 e Capture representative plots with axes properly labeled You can use the Linux screen capture feature or find another way to capture plots into a file such as by using supsimage to make PostScript plots Feel free to use different values of perc and different colormaps than were used in the previous examples The OpenOffice or LibreOffice Word wordprocessing program is an easy program to use for this e
116. enOffice or LibreOffice Writer that you should be able to drag and drop PostScript files into your OpenOffice or LibreOffice Writer document directly When using any graphics program you need to be aware of its options In SU make sure that you type supsimage psimage supswigb pswigb supswigp pswigp so that you can see all of the various options that these programs may expect Note that the dimensions of the plots are in inches and that some programs expect plot dimentions as wbox hbox whereas others expect the dimensions to be given as width height For example Note that the color schemes are completel different in the PostScript generating programs from thei X windows graphics program counterparts 91 7 10 Lab Activity 8 Kirchhoff Migration of Zero offset data In 1978 Bill Schneider published his Kirchhoff migration method The program that implements this in SU is called sukdmig2d There are two shell scripts in data cwpscratch Data2 that you will need to be able to run this on the simple su data Go to your home directory make a temporary directory called Temp2 and copy cd scratch yourusername mkdir Temp2 cp data cwpscratch Data2 Rayt2d simple Temp2 cp data cwpscratch Data2 Kdmig2d simple Temp2 cp data cwpscratch Data2 simple su Temp2 cp data cwpscratch Data2 vel kdmig simple Temp2 cd Temp2 Furthermore you will need to make sure that the headers are correct on simple
117. enient for a particular instrument that is recording data These are usually formats that are dictated by the available storage and the process by which the instruments collects and digitizes the data Such formats often make sense in this usage but may not be easy to work with in computer programs The data may be multiplexed or may be compressed in some other fashion Examples of seismic acquisition formats include SEG D SEG B and SEG 2 Each of these were designed in conjunction with the needs of multichannel seismic acquisition systems SEG D is used for exploration seismic data the other two are small seismograph systems Some data acquisition systems give the user the option of writing out data in the SEGY format However in many cases this is not SEGY by the book but a version that is called the DOS_SEGY format DOS_SEGY is based loosely on the SEGY format but deviates from the official standard 57 Internal formats The term internal may refer to software or to an organization such as a school or a company Internal formats are just that internal Data in such a format is not generally for public consumption or for transport to other systems or exchange but is a format that may make it easier for a particular suite of programs to operate The SU data format is an internal format Every commercial seismic package has its internal format Such systems would include PROMAX DISCO etc Some software packages that specializ
118. epth at source int gwdep Water depth at receiver group short scalel Scalar to be applied to the previous 7 entries to give the real value Scalar 1 10 100 1000 10000 If positive scalar is used as a multiplier if negative scalar is used as a divisor short scalco Scalar to be applied to the next 4 entries to give the real value Scalar 1 10 100 1000 10000 If positive scalar is used as a multiplier if negative scalar is used as a divisor int sx Source coordinate X int sy Source coordinate Y int gx Group coordinate X int gy Group coordinate Y short counit Coordinate units for previous 4 entries and for the 7 entries before scalel Length meters or feet seconds of arc Decimal degrees Degrees minutes seconds DMS e UNBE Il In case 2 the X values are longitude and the Y values are latitude a positive value designates the number of seconds east of Greenwich or north of the equator 51 In case 4 to encode DDDMMSS counit DDD 10 4 MM 10 2 SS with scalco 1 To encode DDDMMSS ss counit DDD 10 6 MM 10 4 SS 10 2 with scalco 100 short short short short short short short wevel Weathering velocity swevel Subweathering velocity sut Uphole time at source in milliseconds gut Uphole time at receiver group in milliseconds
119. ere a scratch directory is provided that also has write permission the user may create his her personal work area via cd scratch mkdir yourusername lt here yourusername is the your user name on the system Unless otherwise stated this text will assume that you are conducting further operations in your personal scratch work area 1 10 Shell environment variables and path The working shell is a program that has a configuration that gives the user access to executable files on the system Recall that echoing the value of the SHELL variable echo SHELL lt returns the value of the users working shell environment tells you what shell program is your working shell environment There are other envi ronmental variables other than SHELL Again note that if this command returns one of the values bin sh bin ksh bin bash bin zsh then you are working in the SH family and need to follow instructions for working with that type of environment If on the other hand the echo SHELL command returns one of the values 19 bin csh bin tcsh then you are working in the CSH family and need to follow the alternate series of in structions given In the modern world of Linux it is quite common for the default shell to be something called binbash an open source version of binsh 1 10 1 The path or PATH Another important variable is the path or PATH The value path variable tells the location that the w
120. ersity We chose the latter and decided that the students should produce as their final project a poster presentation similar to those seen at the SEG annual meeting Terry seemed to think that we could just hand the students the SU User s Manual and the data and let them have at it I felt that more needed to be done to instruct students in the subject of seismic processing while simultaneously introducing them to the Unix operating system shell language programming and of course Seismic Unx In the years that have elapsed my understanding of the subject of seismic processing has continued to grow and in each successive semester I have gathered more examples and figured out how to apply more types of processing techniques to the data My vision of the material is that we are replicating the seismic processors base ex perience such as a professional would have obtained in the petroleum industry in the late 1970s The idea is not to train students in a particular routine of processing but to teach them how to think like geophysicists Because seismic processing techniques are not exclusively used on petroleum industry data the notion of geophysical image processing rather than simply seismic processing is conveyed 10 Chapter 1 Seismic Processing Lab Preliminary issues 1 1 Motivation for the lab In the lecture portion of the course GPGN452 561 now GPGN461 561 Advanced Seis mic Methods Seismic Processing the
121. ethods were that the results should be similar to sonar That is the expectation was that the signal would travel straight down into the subsurface reflect off of a few structures and then travel straight back up to be recorded There was no expectation of off vertical reflections We of course know that this is not the case By the 1930s many geophysicists were well aware of the geometrical issues at the heart of proper seismic interpretation With the formation of the Society of Exploration Geophysicists in 1930 followed by the first issue of the Society s journal Geophysics the proper usage of seismic data for geologic interpretation became known to the geophysical community In Figure 6 1 we see the classical bowtie feature seen over a syncline To the early interpreter of seismic data this diagram would not have constituted an image of the subsurface but rather a source of geometrical data such as dip pertaining the subsurface reflector Another notion that became apparent is that parts of the data on the seismic traces is displaced from its correct position by the properties of wave propagation Assuming that all reflections are normal incidence for this zero offset geometry it is clear that parts of the bowtie originate from higher positions on the sides of the syncline Thus the notion of migrating those arrivals to their correct location became an important idea for interpretation Because the seismic data were analo
122. ever we reverse directions mapping from from b to a or from d to c then we are 78 a b G30 z c x gt gt d x gt t z x z Figure 6 9 Cartoon showing the relationship between types of migration a shows a point in 7 j b the impulse response of the migration operation in x z c shows a diffraction d the diffraction stack as the output point x z modeling or doing data based de migration which is the inverse of migration The idea then is that modeling is the forward process and migration is the inverse operation 6 5 Concluding Remarks The notion of the value and motivation of using seismic data has changed through the history of seismic methods Originally seismic data were used to find an estimate of perhaps only a single reflector As the technique developed the depth to and dip of a specific target reflector was found Most notably was Frank Rieber s dip finder The dip finder was a recording system that was effectively an analog computer that delivered an estimate of depth and dip for stronger reflectors These data were then used for drawing geologic cross sections In fact Frank Rieber s dip finder was doing something similar to a variety of migration called map migration As the petroleum and natural gas industry evolved so did the importance of the seismic method The technique started out as an aid in interpretation becoming later an imaging technology Today
123. find the correct filter range 11 1 5 Lab Activity 19 Spectral whitening of the fake data It may have occurred to the reader after seeing the spectra of the real data that it may be of benefit if the spectra of the traces were flat instead of having the many peaks and valleys This is the process known as spectral whitening We expect that such an operation would tend to sharpen data but with the caveat that we know it will sharpen the noise as well making everything more spike like The basic idea is simple We take the data into the frequency domain and consider the representation of the data d w as a complex valued function d w d w e We then multiply the amplitude d w by a function 1 d w taking care not to divide by zero so that dnew w 1 This may be over the full range of 0 to the Nyquist frequency or over some partial range We then inverse Fourier transform the data This operation is guaranteed to change the relationship between the amplitude and the phase function as 164 trace number b 20 40 60 EE EES Hi eas RESET ELS thetic data multiples fake data spectrally whitened Figure 11 2 a Original fake data b fake data with spectral whitening applied Note that spectral whitening makes the random background noise bigger where d is the called the real part of d and d is the called the imaginary part of d Here the amplitude is given by the modulus
124. following them When you have finished typing in the contents of the Migtest then save the file You will need to change the permissions of this file to give the script execute permission This is done via chmod utx Migtest You may now run the shell script by simply typing Migtest If you get a Migtest command not found error but the permissions are correct you likely need to have on your path You can change your path or type equivalently Migtest 89 to run Migtest Each command within the shell script is run in succession We may consider writing a shell script called ViewMig with the contents bin sh Stolt suxwigb lt stolt simple su title Stolt wbox 800 hbox 200 d2 40 xbox 200 ybox 550 amp gazdag suxwigb lt gaz simple su title Gazdag wbox 800 hbox 200 d2 40 xbox 200 ybox 550 amp phase shift suxwigb lt ps simple su title Phase Shift wbox 800 hbox 200 d2 40 xbox 200 ybox 450 amp finite difference suxwigb lt fd simple su title Finite Difference wbox 800 hbox 200 d2 40 xbox 150 ybox 350 amp Fourier finite difference suxwigb lt ffd simple su title Fourier Finite Difference wbox 800 hbox 200 d2 40 xbox 150 ybox 350 amp split step suxwigb lt split simple su title Split step wbox 800 hbox 200 d2 40 xbox 150 ybox 250 amp phase shift plus interpolation suxwigb lt pspi simple su title PSPI wbhox 800 hbox 200 d2 40 xbox 0 ybox 0
125. g issues and the fact that data must be gathered quickly we really are only interested in relative values 43 Chapter 4 Help features in Seismic Unix Scientific processing and data manipulation packages usually contain many commands Seismic Unix is no exception As with any package there are help features to help you navigate the collection of programs and modules The first thing that you must do with any software package is to locate and learn to use the help features in the package Usually these help mechanisms are not very helpful to the beginner but are really more like quick reference guides for people who are already familiar with the package There are a number of help features in SU here we will discuss only three 4 1 The selfdoc All Seismic Unix programs have the feature that if the name of the program is typed with no arguments a self documentation feature called a selfdoc is listed Try suplane suximage suxwigb sunormalize A A A A For example suplane yields SUPLANE create common offset data file with up to 3 planes suplane optional parameters gt stdout Optional Parameters npl 3 number of planes nt 64 number of time samples 44 ntr 32 number of traces taper 0 no end of plane taper 1 taper planes to zero at the end offset 400 offset dt 0 004 time sample interval in seconds plane 1 dip1 0 dip of plane 1 ms trace leni 3 xntr 4 HORIZONTAL extent of plane traces c
126. g rather than digital such corrections would naturally be applied graphically While graphical migration techniques had been applied since the 1930s the first notable technical paper describing this technique was published by J G Mendel Hage doorn in 1954 This paper is important because Hagedoorn s description of the migration process inspired early digital computer implementations of migration 69 Depth m Time in Seconds 20 40 60 TT ui acl 0 1000 2000 3000 J Synthetic Seismogram v 2000 m s p const v 3000 m s p const Simple Single Reflector Model Figure 6 1 a Synthetic Zero offset data b Simple earth model 70 4000 6 1 Migration as reverse time propagation One way of looking at migration is as a reverse time propagation The idea may be visualized by running the output from a forward modeling demo in reverse time Do the following noting that you need to replace yourusername with your actual username on the system so that the items are copied to your personal scratch area cp data cwpscratch Datail syncline unif2 scratch yourusername Temp1 cp data cwpscratch Datal XSyncline scratch yourusername Temp1 more XSyncline more syncline unif2 Now cd scratch yourusername Temp1 If you type more syncline unif2 0 0 4000 0 1 99999 0 1000 500 1100 1000 1300 2000 1000 2700 1100 3200 1000 4000 1050 1 99999 yo
127. gain jon 1 cdp 1501 2142 su 157 e Then adapt the shell script Radon final located in data cwpscratch Data5 to perform the radon multiple suppression on each of these subsets of the data using the respective tnmo and vnmo This part is time consuming taking several hours for each part so start early Note that the tnmo and vnmo values go in a files with names specified on the line in the shell script Radon final that starts par filename The default filename is radonnmovel par All SU programs have a hidden feature that the commandline argument for parameters can by kept in a file called a parfile and read into the appropriate program via par parfilename e You should have at least one of these subsets processed for class next week Perform the brute stack operation from Homework 5 on that subset For your report show the semblance plot of the single CDP gather after multiple suppression and show the brute stacks of the 500 CDP panel for the data before and after multiple suppression e If you succeed in performing the multiple suppression on all 4 subsets of the data concatenate these together to form the full processed dataset via a command of the form cat radon gain jon 1 cdp 0 500 su radon gain jon 1 cdp 501 1000 su radon gain jon 1 cdp 1001 1500 su radon gain jon 1 cdp 1501 2142 su gt radon gain jon 1 cdp su In this case your report should show the semblance plot of a single CDP gather and show a
128. gration A problem with the original Gazdag phaseshift migration is that it did not handle lateral velocity variation well An approach called phase shift plus interpolation PSPI was developed Jeno Gazdag in 1984 that partially alleviates this problem In SU this is implemented as sumigpspi 87 SUMIGPSPI Gazdag s phase shift plus interpolation migration for zero offset data which can handle the lateral velocity variation sumigpspi lt infile gt outfile vfile optional parameters Required Parameters nz number of depth sapmles dz depth sampling interval vfile name of file containing velocities Please see Notes below concerning the format of vfile Optional Parameters dt from header dt or 004 time sampling interval dx from header d2 or 1 0 midpoint sampling interval Try sumigpspi lt simple su dx 40 dz 12 nz 150 vfile vel fdmig simple gt pspi simple su All of these programs are similar in structure with only the interpolation algorithm being different Each of these algorithms is easily extended to prestack application with amplitudes not being preserved Note All of these programs expect an input velocity in terms of interval velocities as a function of zx z Correspondingly the output files of all of these programs are depth sections which if everything has been done correctly is a representation of a cross section through the earth 7 8 Lab Activity 7 Shell scripts By now everyone
129. hat are not the cross section This is not exactly the cross section of the data sonar su but it gives the idea Rarely are we able to slice into the actual model in this fashion 67 5 7 Homework Problem 2 Time to depth conversion of the sonar su and the radar su data Due Thursday 6 Sept 2012 Find the necessary velocities to permit the correct time to depth conversion of the sonar su and radar su data You will need to figure out the appropriate units because it is not possible for these non seismic datasets to have an accurate representation of the time sampling interval represented in the trace header field dt Make sure that you give a justification explaining why your choice of the appropriate power of 10 scaling factor is likely the correct one Remember that the depth scale on you output data should make sense 5 8 Concluding Remarks When receiving software either that is given to us or that which we purchase it is important to try to figure out what assumptions are built into the software One way to do that is to try the software out on test patterns As applied scientists and engineers we are often in in situations where we are forced to use a tool that is not quite right for the job It is not uncommon for laboratory exper imentalists or ground penetrating radar practitioners to use seismic processing software to do part of the analysis of non seismic data We must be careful to keep the problem simple and expect only w
130. hat on the junk1 su data the picture does not start getting distorted until after about depth 105 This gives a clue as to the place where the faster speeds kick in You will further notice that the junk2 su data does not look very much like the junk su data The first thing that you should notice is that the original junk su data only goes to about 24 seconds but the junk2 su data goes to more than 5 seconds It is a good idea to see if the header values have changed by using surange The original data shows surange lt junk su 32 traces tracl 1 32 1 32 tracr 1 32 1 32 offset 400 ns 64 dt 4000 whereas the depth converted data has a greater number of samples surange lt junk1 su 32 traces tracl 1 32 1 32 tracr 1 32 1 32 trid 30 offset 400 ns 126 lt ns has increased dt 4000 di 3 000000 and finally the depth to time converted data surange lt junk2 su 32 traces tracl 1 32 1 32 tracr 1 32 1 32 offset 400 ns 63 lt ns is now 63 dt 4000 shows ns 63 rather than the original ns 64 samples 64 5 4 1 How time depth and depth time conversion works The way that this works is simple Each sample of the data is a function of time We have velocities to use for each time value If the velocity is constant then the process of time to depth conversion is more of a relabeling process than a calculation However for situations where the velocity varies as we go
131. hat we deserve from the data When receiving data it is important to know everything that you can possibly know about the data such as the spacing of the traces the time sampling interval any pro cessing that has been applied When explaining seismic imaging to non geophysicists it is tempting to say that seis mic imaging is sort of like sonar However sonar is a rough surface scattering imaging whereas seismic is a specular or mirror reflection scattering imaging In rough surface scattering the image may indeed be formed by straight down and straight back re flections In seismic this is rarely the case We must take offset between the source and receiver into account In seismic there are also rough surface contributions These are the diffractions we look for on stacked data In the early days of seismic well into the 1930s there were practitioners of the seismic method who thought that seismic was like sonar and expected that seismic datasets should be data images Such phenomena as bowties provide a clue that seismic is not the same as sonar 68 Chapter 6 Zero offset aka poststack migration The first reflection seismic experiment as applied to the petroleum exploration was con ducted by physicist John Clarence Karcher in Oklahoma in 1921 Oklahoma was the center of the US oil industry at that time It is clear from reading documents from that era that the expectations of reflection seismic m
132. he SU data format is based on the SEG Y format but is not the same So we must convert our data from the data exchange format to the SU format before we can work on the data 9 2 Getting to know our data trace header values Once we have converted the dataset to SU format there are many ways to begin learning about the data For example we might want to merely view the size of the dataset with ls 1 ls l1 seismic su will tell us the size of the data In this case it also is about 800 megabytes We may use surange to determine the header settings in particular to see if they are correct surange lt seismic su 120120 traces tracl 1 120120 1 120120 tracr 1 120120 1 120120 fldr 3 1003 3 1003 tracf 1 120 1 120 ep 101 1112 101 1112 cdp 1 2142 1 2142 cdpt 1 120 1 120 trid 1 112 nhs 1 offset 3237 262 3237 262 gelev 10 selev 6 scalel 1 scalco 1 sx 3237 28512 3237 28512 gx 0 28250 0 28250 counit 3 mute 48 ns 1500 dt 4000 Because this takes awhile to run once you have obtained the output from surange open a file name Notes with your favorite editor and copy and paste the screen output from surange into Notes We can see such information as the range of header values see if the numbers make sense That sort of thing We need to know the total number of traces the total number of shots the number of receivers per gather the time sampling interval th
133. he long offset shallow portion of the section 10 10An average over all of the shots showing direct arrivals head waves wide angle reflections and a curve along with muting may be applied to elimi j te these waves s Toms tasa Sent oe ht ese keel Be bag ee Ae 11 1 a Amplitude spectra of the traces in CMP 265 b Amplitude spectra atter Alterne ey wy ere we See wi See Shoe te dd ME hes Gor eG Yo mS G 11 2 a Original fake data b fake data with spectral whitening applied Note that spectral whitening makes the random background noise bigger 11 3 a Autocorrelation waveforms of the fake su data b Autocorrelation wave forms of the same data after predictive spiking decon 153 Preface I started writing these notes in 2005 to aid in the teaching of a seismic processing lab that is part of the courses Seismic Processing GPGN561 and Advanced Seismic Methods GPGN452 later GPGN461 in the Department of Geophysics Colorado School of Mines Golden CO In October of 2005 Geophysics Department chairman Terry Young asked me if I would be willing to help teach the Seismic Processing Lab This was the year following Ken Larner s retirement Terry was teaching the lecture but decided that the students should have a practical problem to work on The choice was between data shot in the Geophysics Field Camp the previous summer or the Viking Graben dataset which Terry had brought with him from his time at Carnegie Mellon Univ
134. he value of the power of t will be a number 1 lt tpow lt 2 There are other effects There may be a general boosting of some amplitudes in such a way that larger amplitudes are boosted more than smaller amplitudes Thus a global power may be applied One example would be to apply a square root function to the data which would be expressed in sugain by the choice of gpow 5 Finally there may be isolated amplitudes that are just too large These amplitudes may be trucated by clipping the data thus the qclip 95 in the application of sugain above In Jon Claerbout s Imaging the Earth there is a discussion of a sophisticated applica tion of gaining that would be applied in sugain via a choice of parameters that translates into the options tpow 2 gpow 5 qclip 95 The latter qclip refers to clipping on quantiles What is a quantile This is so useful that we have a special parameter for this called jon wherein jon 1 applies this combination of parameters in sugain Caveat Througout this set of notes jon 1 is used because it is convenient not because it is optimal It is up to the you to experiment with sugain to find the optimal gaining It may be that you want to run sugain separately with 119 sugain lt shot ep 200 su tpow 2 gpow 5 qclip 95 suxwigb and try changing the values of tpow gpow and qclip to see the effects of varying these parameters Try to find an optimal setting for the panel One way of testing your c
135. he waveforms in the data will be converted to spikes This is ungapped predictive deconvolution or spiking decon in the parlance of the community We may determine the necessary lag by taking the autocorrelation using suacor and then we will apply the Wiener filter with supef For example consider the fake su data suacor lt fake su ntout 51 suxwigb perc 90 172 trace number trace number a 20 40 60 b 0 054 0 054 2 0 10 4 time s acor time s io 0 154 0 154 autocorrelation waveforms fake data spiked Figure 11 3 a Autocorrelation waveforms of the fake su data b Autocorrelation wave forms of the same data after predictive spiking decon 173 The choice of ntout 51 means that we want 51 samples on the resulting traces This number is chosen to be sufficiently large
136. hell script to execute Migtest fd Migtest ffd Migtest split Migtest pspi Migtest gb Better yet if you want to get timing information you may use the Unix time function time Migtest fd time Migtest ffd time Migtest split time Migtest pspi time Migtest gb Compare your results The notion of cost must be apparent by now Some algorithms are more expensive in computer time than others If we merely want to have quick looks then you cannot beat Stolt migration However as more image quality is desire and more realistic models of the background wavespeed is desired then the more expensive migrations become more attractive 109 8 3 Homework Assignment 4 Due 20 Sept 2012 Migration comparisons Run three of the above migration shell scripts and report on the computational cost and quality of the migrations Show commands run make properly labeled plots and write brief commentary again a maximum of 3 pages in PDF format and email to your instructor Which migration gives the best image Which migration is the fastest 8 4 Concluding Remarks It may have occurred to you to ask why so many different migration algorithms exist In part this is cultural Within company and academic environments different theories of migration were explored Part of this is owing to the business culture If one company is supplying a trademarked service can other companies compete unless they invent their own better
137. hoice is to view your data with suxgraph sugain lt shot ep 200 su tpow 1 0 suxgraph sugain lt shot ep 200 su tpow 1 5 suxgraph sugain lt shot ep 200 su tpow 2 0 suxgraph sugain lt shot ep 200 su tpow 2 5 suxgraph sugain lt shot ep 200 su tpow 1 gpow 5 suxgraph sugain lt shot ep 200 su tpow 1 5 gpow 5 suxgraph sugain lt shot ep 200 su tpow 2 gpow 5 suxgraph sugain lt shot ep 200 su tpow 1 gpow 5 qclip 99 suxgraph sugain lt shot ep 200 su tpow 1 gpow 5 qclip 99 suxgraph labell time s label2 amplitude and so forth When we plot our data in this fashion with suxgraph we are overlaying all of the traces in the gather The resulting plot shows a crude estimate of the envelope of the waves in the gather by plotting all of the traces on top of one another The envelope is a mathematical surface containing the amplitudes but ignoring the oscillation If we are successful in removing the decay in amplitude with time then the amplitude of the envelope will be more or less constant with time The idea is to see which combination of the parameters corrects the geometrical spreading and attenuative decay with time so that the amplitudes in the trace are roughly the same for most of the length of the trace If there are noise spikes or other over corrected spike events we use the qclip to suppress those Note that you probably will not need to set perc on the resulting imag
138. i ae 154 10 3 4 Identifying waves to be muted oa a a a 2 004 154 10 3 5 How to pick mute values Soc ee ah a oe es es ee 155 10 3 6 The shape of the wavelet o a a a a a ee ee 156 10 3 7 Further processing lt e c etu ae oe ei a a a E ALE a a a a aia 156 10 4 Homework Assignment 7 due Thursday 11 Oct 2012 before 9 00 AM 157 10 4 1 The at command using the computer while you are asleep 158 10 5 Concluding remarks oaoa aa Me ale OMG MON hl ie Bd 160 Spectral methods and advanced gaining methods for seismic data pro cessing 161 11 1 Common assumptions of spectral method processing 161 TE COBUS AMIN y wren Dd sete Adee een amp det As ce oe eid Meck Ee og 162 TREIZ Wiad phases e t ccentla eo aine tht a0 Bb ee eta oleae Dre okay ew we 162 11 1 3 White spectrum fe Sbload alah he ete eee deg 163 11 1 4 Lab Activity 18 Frequency filtering 164 11 1 5 Lab Activity 19 Spectral whitening of the fake data 164 11 1 6 Spiking deconvolution also known as minimum phase deconvolu tion or prediction error filtering ace AS a ewe A 166 11 2 An overview of deconvolution mathematics 167 11 2 1 Convolution of a wavelet with a reflectivity series 167 11 2 2 Convolution with a wavelet o o oao a a 169 14233 Deconvolution 2 niad ea a a ee ee a A ee 169 11 2 4 Division im the frequency domain 44 4 ska keke es 170 11 2 5 Cross and auto correlation o
139. ifted both temporally and spatially suggesting that the filter doing the shifting must be a filter that operates in both the frequency and wavenumber domain If the velocity function is variable with time the Stolt method accounts for this by applying a stretch to the data prior to the filtering operation To get the shifting correct Robert Stolt based his shifting filter on an integral equation representation of the wave equation and made use of the fast Fourier transform algorithm for speed To handle variable wavespeed Stolt introduced a stretch much as we scaled the time section to appear interchangeable with depth 7 2 1 Stolt migration of the Simple model data Here apply Stolt migration to Copy the dataset into your local working directory via cd scratch yourusername mkdir Temp2 pwd you should be in scratch yourusername cp data cwpscratch Data2 simple su Temp2 changing your working directory to Temp2 via cd Temp2 You may view the data via suxwigb lt simple su xcur 3 title Simple The Stolt migration is performed via the program sustolt To see the self documentation for sustolt type sustolt SUSTOLT Stolt migration for stacked data or common offset gathers 82 sustolt lt stdin gt stdout cdpmin cdpmax dxcdp noffmix Required Parameters cdpmin minimum cdp integer number for which to apply DMO cdpmax maximum cdp integer number for which to apply DMO dxcdp dist
140. ild a collection of velocities chosen from the well log and assign the times from the arrival time on the seismic section and thus we would define a v t If we have v t for specific locations in the rock volume we are investigating we could assemble a u t x or a u t x y through some sort of interpolation 7 1 1 Velocity conversion v t to v z The conversion between these two is given by the Dix equation 2 2 1 2 toUbms2 Ups t2 ti Vint 80 where Vint is the interval velocity t is the traveltime to the first reflector t is the traveltime to the second reflector Upms is the root mean squared velocity of the first reflector and U pms2 is the root mean squared velocity of the second reflector The RMS root mean squared part comes fromt the fact that we are effectively averaging and taking the square root of squares going from RMS to interval velocity Often the output of a migration is horizontal position versus migrated time This implies that the input velocity is expressed as v t rather than v x z We may have interval or rms velocities as a function of time e g Vinz t Or Upms t This may seem strange at first but if we have v z profile then v t is obtained by converting time to depth In SU a program called velconv performs many simple conversions between interval and rms velocities Type velconv with no options to see the selfdoc of the program velconv VELCONV VELocity CONVersion
141. into a depth scale that is compatible with the horizontal scale If we draw a circle centered at time t 0 of a given seismic trace and passing through a given seismic arrival we have sketched all possible reflection points from which the seismic arrival could have originated These circles are the same as the incident field seen in the seismic movie If we recall seismic migration finds the place where the incident field interacts with the reflected field the reflector surface When similar circles are drawn for every arrival on every trace the result is a collection of circles whose envelope delineates the reflector See Fig 6 3 for an idea of what this should look like 73 0 Time in Seconds Synthetic Seismogram Figure 6 2 The Hagedoorn method applied to the arrivals on a single seismic trace Trace 20 40 60 8 Figure 6 3 Hagedoorn s method applied to the simple data of Fig 6 1 Here circles each centered at time t 0 on a specific trace pass through the maximum amplitudes on each arrival on each trace The circle represents the locus of possible reflection points in x z where the signal in time could have originated 74 Trace 20 40 60 80 0 xe Cc O Q 35 if UT ey es ctl 2 Synthetic Seismogram Figure 6 4 The dashed line is the interpreted reflector taken to be the envelope of the circles Mathematically this method of migration may be thought of as the reconstruction of th
142. ion Every wavefront for both positive and negative time t is found by passing a plane parallel to the x z plane through the cone at the desired time t We may want to run time backwards for migration The light cone representation for negative times is now embedded in the x z t cube A seismic arrival to be migrated at the coordinates T is placed at the apex of the cone The circle that we draw on the seismogram for that point is the set of points obtained by the intersection of the cone witi OE AMG nzr Ve lee ate aa se Sede bce cat eines eee Ee Sth a Hagedoorn s method of graphical migration applied to the diffraction from a point scatterer Only a few of the Hagedoorn circles are drawn here but the reader should be aware that any Hagedoorn circle through a diffraction event will intersect the apex of the diffraction hyperbola The light cone for a point scatterer at z z By classical geometry a vertical slice through the cone in x t the z 0 plane where we record our data is a hyperbola Time migrations collapse diffraction hyperbolae to their respective apex points Depth migrations map these apex points int the Go per a 2D 0 2 tc ame ee me a aa a Na ee a i Cartoon showing the relationship between types of migration a shows a point in 7 j b the impulse response of the migration operation in x z c shows a diffraction d the diffraction stack as the output point FN vast ese obviates he de ON afte s
143. ipt Try to find better tnmo and vnmo values than the ones that are in the script to suppress the multiples e Show suxwigb or supswigb plots of the single CDP before and after multiple suppression and show the semblance plot of the data after multiple suppression Because this is a warmup assignment for more complicated applications later on here are a few tips and tricks 1 Make sure you are working with the shell script Test located in data cwpscratch Data5 not the shell script Radon test discussed in the previous section Test has some tnmo vnmo values already set for you to get you started Remember that we are using the NMO operator as a tool to separate the moveouts of arrivals that we want to get rid of the multiples from the arrivals we want to keep the reflections Every application of a forward NMO is followed by an application of filtering in the radon domain which in turn is followed by an application of Inverse NMO to undo the original NMO 150 The goal of velocity analysis is to get the correct stacking velocities In our case these values will give us two benefits We can use these tnamo and vnmo for the preprocess for the radon transform domain filtering and ultimately we will get the NMO velocities to stack the data The idea is to eliminate the multiples by filtering in the radon domain and then pick new vnmo and tnmo from your semblance plot How to pick You can just zoom in on the se
144. is application of prediction error filtering in the frequency domain called fx decon that was created in the 1984 by L L Canales This technique uses predictive decon in the space frequency domain to identify and eliminate random noise For example consider the spectrally whitened version of fake su suwfft w0 0 wi 1 w2 0 lt fake su suifft gt white fake su Applying sufxdecon to these data sufxdecon lt white fake su suxwigb perc 99 Try this operation on different versions of CDP 265 It may be best to reserve FX decon for the later stages of processing after the stack 11 6 Lab Activity 20 Wavelet shaping Many papers written in the 1970s dealt with the issue of wavelet estimation That is using statistical methods to determine the shape of the average wavelet throughout the dataset or in regions in a dataset The motivation for this is to use deconvolution to change the waveforms of the data to a new desired output waveform Currently in SU there is no sophisticated wavelet estimation code as yet The user can get a crude estimate of the wavelet by selecting the waveform from a horizontal portion of a reflector in the data Knowing the trace number and the time window of the wavelet we may use suwind to capture this average wavelet via suwind key tracl min TRACE max TRACE tmin TMIN tmax TMAX gt wavelet su 177 where TRACE TMIN and TMAX are replaced with the actual values of the trace number and
145. is tired of typing the same commandlines repetetively Furthermore it is easy to forget exactly what was typed for further testing To remedy this problem we will now capture these commandlines into a program called a shell script The shell script is one of the most powerful aspects of Unix and Unix like operating systems For example we can capture all of the migrations above into a single script for com parison Begin by typing cd Temp2 and opening a file called Migtest using your favorite editor The contents of Migtest should be 88 bin sh set xX Stolt sustolt lt simple su cdpmin 1 cdpmax 80 dxcdp 40 vmig 2000 tmig 0 0 gt stolt simple su gazdag sugazmig lt simple su dx 40 vmig 2000 tmig 0 0 gt gaz simple su phase shift sumigps lt simple su dx 40 vmig 2000 tmig 0 0 gt ps simple su finite difference sumigfd lt simple su dx 40 dz 12 nz 150 vfile vel fdmig simple gt fd simple su split step sumigsplit lt simple su dx 40 dz 12 nz 150 vfile vel fdmig simple gt split simple su phase shift plus interpolation sumigpspi lt simple su dx 40 dz 12 nz 150 vfile vel fdmig simple gt pspi simple su exit 0 The top line indicates that the Bourne shell sh is being used to run the commands that follow The set x tells the Bourne shell interpretor to echo each command as it is being run The backslashes indicate are line continuation symbols and should have no spaces
146. ity for time ranges you see frowns in the data increase the velocity and perform Stolt migration again Repeat this iteration of migration followed by velocity modification until you are satisfied that you have a good set of velocities Use suintvel to convert your RMS velocities into interval velocities Using dz 3 nz 1500 with these velocities use suttoz to make an approximate depth section out of your Stolt migrated image We performed some of these operations in the early part of the semester You may consult the earlier chapters of the notes for details Show your Stolt time section and your Stolt depth section and list your velocity time pairs Don t be afraid to use small clip values to accentuate weaker arrivals that you see in the section Discuss what you see Additional tips For students who wish to redo their multiple suppression please make use of the shell script Velan radon located in data cwpscratch Data5 This shell script is a version 190 of the Velan script set up to allow you to make picks explicitly for multiple suppression You may also want to apply predictive deconvolution to suppress the near offset multiples before applying Velan radon and Radon final The dz 3 and nz 1500 in suttoz is the source of the 4500 you see for the maximum depth on your depth section With d1 8 in the headers on the output to suximage we are showing the data to a depth of 4500 meters The question for the processor i
147. ity profile it is clear that we need to both suppress multiples and perform velocity analysis so that we can get a better stack 9 8 Homework 5 Due Thursday 27 Sept 2012 prior to 9 00AM Repeat the gaining and raw and brute stack operations of the previous sections exper imenting with sugain and the values of tpow gpow and qclip to find gaining parameters that you believe work better to balance the amplitudes of the data in time Replace the jon 1 in the examples below with your parameters To repreat the brute stacking operations put in the additional windowing step suwind key offset min 3237 max 1000 to pass only the far offset data For example sunmo vnmo 1500 lt gain jon 1 cdp su suwind key offset min 3237 max 1000 sustack gt stack far gain jon 1 cdp su Similarly stack only the near offset data for example sunmo vnmo 1500 lt gain jon 1 cdp su suwind key offset min 1000 max 262 sustack gt stack near gain jon 1 cdp su and compare these results Perform a similar test for each of the NMO velocity cases in the previous section Do we really want to include the near offset traces when we stack What is on these traces and why does nmo 1500 accentuate this part of the dataset 9 8 1 Are we done with gaining The issue of amplitude corrections are complicated The example in the Homework assignment above is really a preliminary operation We can see this by asking what processes should still be
148. ive apex points Depth migrations map these apex points into the x z 2D plane depth section but could be a time section A useful diagram for understanding the diffraction stack is the light cone diagram in Figure 6 8 A light cone is the representation of the surface where solutions of the wave equation live The scatterer is located at the point x z Time increases downward A horizontal slice through the cone reveals the circular wavefronts that are the circles drawn in Hagedoorn s method A vertical slice through the cone in x t reveals the hyperbola that is the characteristic shape of a diffraction in a constant wavespeed medium 6 4 Migration as a mathematical mapping Another diagram that reveals migration as a type of data transformation or mapping may be seen in Figure 6 9 Here we see that the impulse response of the migration operator is a circular curve in constant wavespeed media The diffraction in x t may be thought of as the impulse response of the modeling operation that made the data from a point at x z Migration by diffraction stack therefore consists of selecting a point x z modeling the diffraction curve in x t and then summing through the data over this curve Note that this must be done for every output point to make the image Figure 6 9 represents more than migration Going from a to b is Hagedoorn s migration method Going from c to d is the diffraction stack migration method If how
149. jem sys yYFnoy UdAd SAY OJUTZ JOY 99S YS SAY JOY SISH JS UAM MON oyuizauqqeal n N po 1940 P NOM sys UNOD OJUIZA Joy UO L 798 0 SJURM PUP PAS UO JUNO JOY OFT posSo St Wqqey eorssor JI o durexa 104 auipusasn JO OUIeUIOSN umo INOA aINIDSQns pur daaoge aes NOA sv ysnf uoNeziyeydes oy asn NOA amns og dS auipusasnjoyu po OJUIZY auDUsaSN OYN po PAG AUDUsISN QIN PI OUIZ AUIDULASN YN Po Aleqoo aumpusasnjayN Po PWOIYD auDUsasnjnyN po TWIA SAS IIO AUB WOI A1OJALIP WAS S AUP 0 JS 0 DSN 0 SPUPLULIOD OY AIL IH SIAO y Jo Aue uo soft INO 0 193 ueo NOA our Passo aw NOA s1oyNdUOd s JO YOTYM ILU ON squnoooe MLY NOA YOTYA UO S SOY XUN BUSS dU 10 PIWLU SaLIOJALIP UI YSIp IUO UO Ajyear ose S H INOA yey SULU WYL SAN PAF AH JIOMION OY UO paS ale JS SMAV LS pue ozur PAS UZ eq WYJ sIomduio xun Teu SLIN A UO poses s l s li SIN Yum uHoM ty References Ivers 29 UNIX Quick Reference card p1 From the Un Figure 2 3 4 JSHASL uv q poeu asou Jdaoxe pueUILIOD YORE JO pu dU W NAA LAA SSAA JAYS Wodf nxa u qof punossasof aumnsay u qof punossyong aunsay PUunos8yobg Ul puvunuos uny u qof punossyovg puadsng ssaoid Juasand puadsng u ssaooid aaoway SIDIS SNIDIS SSaIOAd JUL u qol m3 sqof Jo 181 Mie spuodas u aof daaqy dmo uaasos aumnsay 8U1 04IS uaas9s dojg sassazoid Jdnssaquy uondrnsq 3x9 u 33 4 3q WPUNUWOI u dos
150. k su We may now apply suradon to transform the data into 7T p domain suradon lt junk su offref 3237 interoff 262 igopt 2 choose 0 pmin 2000 pmax 2000 dp 8 depthref 1000 suximage perc 99 label1 tau label2 p Negative values of p correspond to upward curving events while p O is anything that is flat Anything curving down which is to say having positive moveout in other words arrivals that are slower than the water speed are to the left The program suradon is a sophisticated improvement on the traditional tau p transform One improvement is that the p values are given as times in milliseconds on the data instead of velocities Also there are several choices of Radon transform that the user may apply Here in igopt 2 mode it sums over hyperbolae rather than mere lines The p value then is the takeoff angle of a hyperbola Exactly matching a hyperbola would tend to make a dot at a particular T p pair on the plot This is an idealized situation We don t have the exact hyperbolae We can control the shape of the hyperbolae with the depthref parameter to some degree What we get are two regimes of curving arrivals When we flatten our data with a NMO correction items with positive p are multiples these we seek to remove Items that are flattened or have negative p are our our data which we want to keep The suradon has a second mode useful as filter for multiple suppression We may select choose 1 to suppress multiples s
151. large spike then it is possible that most of the 256 shades are used up by that one amplitude Therefore scaling amplitudes is often necessary The simplest processing of the data is to amplitude truncate clip the data The term clip refers to old time strip chart records which when amplitudes were too large appeared if someone had taken scissors and clipped of the tops of the sinusoids of the oscillations Try suximage lt sonar su perc 99 amp suximage lt sonar su perc 99 legend 1 The perc 99 passes only those items of the 99th percentile and below in amplitude You may need to look up percentile on the Internet In other words it clips amplitude truncates the data to remove the top 1 per cent of amplitudes Try different values of perc to see what this does 33 200 400 0 05 0 10 0 15 0 20 0 25 Figure 3 1 Image of sonar su data no perc Only the largest amplitudes are visible 34 200 400 0 05 0 10 0 15 Figure 3 2 Image of sonar su data with perc 99 Clipping the top 1 percentile of amplitudes brings up the lower amplitude amplitudes of the plot 35 3 3 Legend making grayscale values scientifically meaningful To be scientifically useful which is to say quantitative we need to be able to translate shades of gray into numerical values This is done via a gray scale or legend A legend is a scale or other device that allows us t
152. ling with 2 way traveltimes Thus for our seismic datasets we would expect the range of k values to be Dies ira lmaz lt k lt where wWmin and wWmar are the minimum and maximum frequencies in the data We freely trade time t for depth x3 so it should not be a shock that we may consider the data not in the f k domain but rather in the 2D wavenumber domain k1 k2 where k is the vertical wavenumber and k is the horizontal wavenumber Figures 7 2c and d show the corresponding images that we obtain by making these assumptions using the program suspeck1k2 to calculate the 2D spatial transform amplitude spectrum The spectrum is symmetric because the data are real valued The Fourier transform of a real valued function is always symmetric 97 wavenumber k_2 wavenumber k_2 0 20 404 60 4 frequency Hz frequency Hz 80 4 80 4 1004 1004 120 120 Simple Simple interpolated c k2 d 2 I k 0 5 0 0 5 1 0 i 0 5 0 0 5 0 54 0 54 Simple k1 k2 domain Simple interpolated k1 k2 domain Figure 7 2 a Simple data in the f k domain b Interpolated simple data in the f k domain c Simple data represented in the kz kr domain d Interpolated simple data in the k k domain The simple su data are truncated in the frequency domain with the aliased portions folded over to lower wavenumbers The interpolated data are not folded
153. lope index 40 60 80 120 7 000 500 0 500 1000 0 2 0 2 0 4 0 4 O O E 0 6 0 6 0 8 0 8 1 0 1 0 Suplane Data Suplane data Radon domain Figure 10 3 a Suplane data b its Radon transform Note that a linear Radon transform has isolated the three dipping lines as three points in the r p domain Note that the fact that these lines terminate sharply causes 4 tails on each point in the Radon domain 139 suradon lt suplanedata su igopt 3 interp 4 choose 0 depthref 1000 interoff 0 offref 1190 pmin 1000 pmax 1000 gt radon su suximage lt radon su perc 99 label1 tau s label2 slope index title suplane data Radon transformed Here it should be noted that suradon is a complicated program with a lot of options so we will approach the problem of using this program cautiously In this case we use choose 0 to get a forward Radon transform of the data igopt 3 to select a linear Radon transform meaning that the curves that are summed over are straight lines The rest of the values make sense if we do surange lt suplanedata su 120 traces tracl 1 120 1 120 tracr 1 120 1 120 offset 0 1190 0 1190 ns 256 dt 4000 which shows that the offset ranges between 0 and 1190 meters The pmin and pmax parameters must be chosen to be large enough to encompass all of the slopes as measured by maximum times to be seen Figure 10 4 shows the result of performing the Radon transform on the data from suplane Beca
154. lue of trace amplitude for each window and normal izes the data within the respective window the by dividing by the sum AGC is roughly data driven but it is somewhat dangerous to use in that the AGC function can lose small or large amplitudes and can introduce artifacts that have more to do with the window size you use and less with the real amplitude decay in the data There are varying opinions as when to and when not to use AGC 9 5 4 Model based divergence correction There is a more sophisticated approach to gaining data which is to model the actual geometrical spreading amplitudes by solving the wave equation for the amplitudes using a velocity model and then normalizing the data based on these calculated amplitude values In SU the programs sudivcor sudipdivcor In modern migration programs it may be that we don t want to gain the data but that the gaining is part of the inverse process that is being applied to the data The correction for geometrical spreading then is built in to the migration process 9 6 Getting to know our data Different Sorting Geometries We need not live with our data in the form of shot gathers By now the reader is aware of the CMP NMO Stack procedure The data are recorded as shot gathers and are resorted to CMP gathers We may sort data in terms of offset to make common offset gathers or by receivers to make receiver gathers or by any other parameter that we might want 9 6 1 Lab A
155. m make sure that the frequency content of the desired output waveform is roughly the same as that of the data so that values not be manufactured by the program 11 7 Advanced gaining operations Before gaining our data we would like to remove the effect of the differing source strengths and receiver gains on our data These effects tend to cause vertical strip ing in our data Indeed this section should probably appear in the section on gaining but as this requires some additional sorting of the data we discuss the operation here We must use some estimate for source strength but we also know that there are likely variabilities due to the receiver gains so a statistical approach is used Here we apply RMS power balancing The approach we use here is fast and simple but it is not the only approach that may be applied 179 11 7 1 Correcting for differing source strengths We first remove source strength by beginning with our data as shot gathers Here this is the file seismic su before any other processing There is a program called susplit which will split the data out into separate files based on header field value To make separate files of shots we first move the data into a convenient location mkdir Temp mv seismic su Temp cd Temp susplit key ep close 1 lt seismic su There are 1001 shots so there will be 1001 files that begin with the word split We may loop over these performing a gaining operation that
156. m 0 to Nyquist 119 Phase of complex trace from 0 to Nyquist 121 Wavenumber time domain k t 49 122 Wavenumber frequency k omega 123 Envelope of the complex time trace 124 Phase of the complex time trace 125 Frequency of the complex time trace 130 Depth Range z x traces 143 Seismic Data Vertical Component 144 Seismic Data Horizontal Component 1 145 Seismic Data Horizontal Component 2 146 Seismic Data Radial Component 147 Seismic Data Transverse Component 201 Seismic data packed to bytes by supack1 202 Seismic data packed to 2 bytes by supack2 short nvs Number of vertically summed traces yielding this trace 1 is one trace 2 is two summed traces etc short nhs Number of horizontally summed traces yielding this trace 1 is one trace 2 is two summed traces etc short duse Data use 1 Production 2 Test int offset Distance from the center of the source point to the center of the receiver group negative if opposite to direction in which the line was shot int gelev Receiver group elevation from sea level all elevations above the Vertical datum are positive and below are negative int selev Surface elevation at source int sdepth Source depth below surface a positive number 50 int gdel Datum elevation at receiver group int sdel Datum elevation at source int swdep Water d
157. mblance plot and read values of the axes Alternatively you may pick values by placing the cursor on the place on the plot that you want to pick Then type the letter s and the value of a time velocity pair will be printed in your terminal window Then edit the shell script to use the new values and then run Test again It is helpful to have a wiggle trace plot of your CDP gather on the screen next to your semblance plot so that you can see what event goes with a given semblance peak Is it a multiple or is it a real reflector that is making a given semblance peak Update the tnmo and vnmo pairs and run Test again to see if the semblance plot shows fewer multiples Repeat until it looks more like a textbook example of a semblance plot What to pick If the NMO velocity is perfect at a particular time were there is a reflector then the semblance plot will have a peak value at that NMO velocity and that time How to know what values to pick Imagine the geology The water speed is 1500 m s at t 0 0 always The next peak will be at the top of the unconsolidated material in the water bottom it will be about 5 s and will be only slightly higher than 1500 m s We expect that as we go down in depth the velocity will increase until we are in consolidated material where it will be higher Speed generally increases with depth though there may be local velocity reductions hard to see on semblance Semblance peaks occur where there are
158. me combination of these with the computer name and or the working directory and or the commandline number echo SHELL lt returns the value of the users j working shell environment type this dollar sign The command echo SHELL tells your working shell to return the value that denotes your working shell environment In English this command might be translated as print the value of the variable SHELL In this context the dollar sign in front of SHELL should be translated as value of Common possible shells are bin sh lt the Bourne Shell bin bash lt the Bourne again Shell bin ksh lt K shell bin zsh lt Z shell bin csh lt C shell bin tcsh lt T shell The environments sh bash ksh and zsh are similar We will call these the sh family The environments csh and tcsh are similar to each other but have many differences from the sh family We refer to csh and tcsh as the csh family 1 6 Setting the working environment Each of these programs have a specific syntax which can be quite complicated Each is a language that allows the user to write programs called shell scripts Thus Unix like systems have scripting languages as their basic interface environment This endows Unix like operating systems with vastly more flexibility and power than other operating systems you may have encountered With more flexibility and power there comes more complexity It is possible to
159. minimum and maximum times that where the wavelet of choice is located Also we can make a desired output waveform by using the program suwaveform SUWAVEFORM generate a seismic wavelet suwaveform lt stdin gt stdout optional parameters Required parameters one of the optional parameters listed below Optional parameters type akb wavelet type akb AKB wavelet defined by max frequency fpeak berlage Berlage wavelet gauss Gaussian wavelet defined by frequency fpeak gaussd Gaussian first derivative wavelet ricker1 Ricker wavelet defined by frequency fpeak ricker2 Ricker wavelet defined by half and period spike spike wavelet shifted by time tspike unit unit wavelet i e amplitude 1 const dt 0 004 time sampling interval in seconds ns if set number of samples in output trace fpeak 20 0 peak frequency of a Berlage Ricker or Gaussian For example suwaveform gt dfile su type rickerl fpeak 15 where dfile su contains a Ricker wavelet with peak frequency fpeak of 15Hz The frequency content of the desired output waveform should approximately match the fre quency content of the input wavelet Given the wavelet wavelet su and the desired output waveform dfile su we may use the wavelet shaping code called sushape SUSHAPE Wiener shaping filter sushape lt stdin gt stdout optional parameters Required parameters w vector of input wavelet to be shaped or yO cease 178 wfile file containing in
160. mo curve up wheres multiple energy 137 10 0 1 Creative use of NMO and Inverse NMO In the sections that follow we will find that the process of NMO correction may be used as a tool to change the slope of our data We also will make use of inverse NMO It is possible to make an approximate inverse of the normal moveout correction Thus we will find that we will apply a sequence of forward NMO followed by a moveout based filtering technique followed by an inverse NMO It is important for the reader to realize that these usages are not the application of NMO for the final flattening of the data Even though sunmo may appear as part of the processing sequence the end result is that there is no net NMO correction applied to the data 10 1 The Radon or 7 p Transform We may exploit the differential moveout in the data between multiples and reflectors by applying an NMO correction to flatten arrivals traveling at the water speed The Radon or T p tau p transform maps the data into the traveltime slowness domain The slowness may be thought of as the slope of the data in time and offset In other words it is a quantity with units of time distance Speed is distance time hence the origin of the term slowness The Radon transform operates by considering each traveltime depth in a seismic dataset and by scanning over the data along curves of different initial slopes or slownesses referenced between the right and left sides of th
161. mpare how the diffractions that appear in the stacked data are changed in the migrated sections Are there artifacts What does the shape of the artifacts tell you about the migration wavespeed Look for geologic structures in the time migrated data Compare these to the time section Compare the depth migrated section to the time migrated section Do you see all of the data in the depth section Look for geologic structures in the depth section Where is the water bottom Do you see an unconformity Do you see any faults Artifacts Horst and graben structures Any suspicious horizons that might be multi ples Is the migration velocity correct Too high Too low 107 8 1 2 Phase Shift migration We may now run the phase shift migration demo script PSmig by typing PSmig You will first notice first that it takes a bit longer to run the phase shift program than it did with Stolt The advantage to phase shift migration is that propagates the field locally down in depth so may handle the local variation of the background wavespeed better The output of this program is a time section You may convert this to a depth section by typing Suttoz psmig As with the Stolt migration example you may view the results via suximage lt seismic3 su clip 2 title Stacked data amp suximage lt ps seis su clip 2 title PS time section amp suximage lt ps depth seis su clip 2 title PS depth section amp 8 1 3 Questions for discussion
162. n sign lt is called redirect in and suxwigb lt junk su amp says run suxwigb reading the input from the file junk su run in background e pipe from program to program 26 a trace number b trace number 10 20 30 5 10 20 30 20 0 05 J 0 10 pil 605 time s ne or an ew anal eer nal Freq Hz 0 154 0 204 1004 1205 0 254 suplane test pattern Figure 2 2 a The suplane test pattern b the Fourier transform time to frequency of the suplane test pattern via suspecfx e gt write data from program to file redirect out e lt read data from file to program redirect in e amp run program in background 2 2 Stringing commands together We may string together programs via pipes and input and output via redirects gt and lt An example is to use the program suspecfx to look at the amplitude spectrum of the traces in data made with suplane suplane suspecfx suxwigb amp make suplane data find the amplitude spectrum plot as wiggle traces Equivalently we may do suplane gt junk su make suplane data write to a file suspecfx lt junk su gt junki su find the amplitude spectrum write to a file suxwigb lt junki su
163. n be imported into documents In SU we have graphics programs that write output in the PostScript language supswigb lt junk su title suplane test pattern labell time s label2 trace number gt suplane eps 2 1 Pipe redirect in lt redirect out gt and run in background amp In the commands in the last section we used three symbols that allow files and programs to send data to each other and to send data between programs The vertical bar is called a pipe on all Unix like systems Output sent to standard out may be piped from one program to another program as was done in the example of suplane suxwigb amp which in English may be translated as run suplane pipe output to the program suxwigb where the amp says run all commands on this line in background The pipe is a memory buffer with a read from standard input for an input and a write to standard output for an output You can think of this as a kind of plumbing A stream of data much like a stream of water is flowing from the program suplane to the program suxwigb The greater than sign gt is called redirect out and suplane gt junk su says run suplane writing output to the file junk su The gt is a buffer which reads from standard input and writes to the file whose name is supplied to the right of the symbol Think of this as data pouring out of the program suplane into the file junk su The lest tha
164. n is to reduce wave amplitudes expo nentially as a function of the number of cycles that the wave has traveled distance in wavelegths and is usually expressed in terms of a quality factor Q For example in the frequency domain a decaying solution may be written as wr ulz y z w w w e 2 and w w is the frequency domain representation of the waveform represented by W t r c above To correct for geometric spreading and attenuative amplitude loss we may apply an amplitude correction known as a gain to the data There are many gaining strategies we will discuss a couple of the more common ones 9 5 2 Lab Activity 12 Gaining the data One way to do this is to multiply the data by a power of time This is done via sugain sugain lt shot ep 200 su tpow 1 gt gain tpow 1 ep 200 su sugain lt shot ep 200 su tpow 2 gt gain tpow 2 ep 200 su and the effect is to multiply each amplitude on the trace by a factor of t A simple way of looking at this is that for an average constant velocity of c the two way traveltime is t 2r c where r is the distance the wave has traveled to the reflector Hence 1 t x 1 r and thus 1 r geometrical spreading balanced by multiplying data by t This does not seem to be quite enough owing to the fact that the wavespeed generally increases with depth and the presence of anelastic attenuation Commonly a factor of t is a better choice but this may be too much It may be that t
165. n the output There is insufficient coverage 83 7 3 Gazdag or Phase shift migration In 1978 Jeno Gazdag published a type of migration that also makes use of the shift theorem In this class of method the model is discretized in such a way that the vertical direction z is considered to be preferred The data are Fourier transformed and the migration is applied as a phaseshift to each wavenumber The data are then inverse Fourier transformed The program sugazmig sugazmig SUGAZMIG SU version of Jeno GAZDAG s phase shift migration for zero offset data with attenuation Q sugazmig lt infile gt outfile vfile optional parameters Optional Parameters dt from header dt or 004 time sampling interval dx from header d2 or 1 0 midpoint sampling interval ft 0 0 first time sample ntau nt from data number of migrated time samples dtau dt from header migrated time sampling interval ftau ft first migrated time sample tmig 0 0 times corresponding to interval velocities in vmig vmig 1500 0 interval velocities corresponding to times in tmig Try the Gazdag migration of the simple su data via sugazmig lt simple su dx 40 vmig 2000 tmig 0 0 gt gaz simple su Note the velocities vmig here are interval velocities as a function of time You will notice that you could do several Stolt migrations in the time that it takes to do a single Gazdag By the application of a slightly different formulation of phase shift migr
166. nds pwd print working directory Is list contents and cd change directory Locating yourself on the system If you type cd pwd 1s 15 Vi Quick Reference http www sfu ca yzhang linux MOVEMENT lines ends at lt CR gt sentence ends at puncuation space section ends at lt EOF gt By Character Marking Position on Screen h oe 1 lt o OE EEE mp mark current position as p a z p move to mark position p move to first non whitespace on line w mark p Miscellaneous Movement j fm forward to character m Fm backward to character m By Line tm forward to character before m nG to line n Tm backward to character after m 0 first last position on line w move to next word stops at puncuation or first non whitespace char on line WwW move to next word skips punctuation first character on next prev line b move to previous word stops at punctuation B move to previous word skips punctuation By Screen e end of word puncuation not part of word F B scroll foward back one full screen E end of word punctuation part of word D U scroll forward back half a screen s next previous sentence AE Y show one more line at bottom top J next previous section L go to the bottom of the screen hf next previous paragraph Zl position line with cursor at top goto matching parenthesis Z position line with cursor at middle Z position line with cursor at EDITING
167. nix like operating system and is available as full source code Students are free to install Linux and SU on their PCs or use Unix like alternatives and thus have the software as well as the data provided for the course for home use during and beyond the time of the course 1 2 1 Steep learning curve The disadvantage that most beginning Unix users face is a steep learning curve owing to the myriad commands that comprise Unix and other Unix like operating systems The advantages of software portability and flexibility of applications as well as superior networking capability however makes Unix more attractive to industry than Microsoft based systems for these expert level applications While a user in an industrial envi ronment may have a Microsoft based PC on his or her desk the more computationally intensive processing work is done on a Unix based system The largest of these are clusters composed of multi core multiprocessor PC systems It is not uncommon these days for such systems to have several thousand cores which is to say subprocessors available Because a course in seismic processing is of broad interest and may draw students with varied backgrounds and varied familiarity with computing systems we begin with the basics The reader familiar with these topics may skip to the next chapter 1 3 Logging in As with most computer systems there is a prompt usually containing the word login or the word username
168. nvention 9 6 4 Viewing the headers We now have the data gained and sorted into CMP gathers We use the old term CDP here because this term is the one SU uses to designate the CMP header field in the SEG Y header As before we can view some header fields in the data sugethw lt gain jon 1 cdp su sx gx offset ep cdp more sx 3237 gx 0 offset 3237 ep 101 cdp 1 sx 3237 gx 25 offset 3212 ep 101 cdp 2 Sx 3262 gx 25 offset 3237 ep 102 cdp 3 Sx 3237 gx 50 offset 3187 ep 101 cdp 3 Sx 3262 gx 50 offset 3212 ep 102 cdp 4 Sx 3237 gx 75 offset 3162 ep 101 cdp 4 sx 3287 gx 50 offset 3237 ep 103 cdp 5 125 offset meters 1000 500 Figure 9 5 A stacking chart is merely a plot of the header CDP field versus the offset field Note white stripes indicating missing shots 126 Sx 3262 gx 75 offset 3187 ep 102 cdp 5 Sx 3237 gx 100 offset 3137 ep 101 cdp 5 sx 3287 gx 75 offset 3212 ep 103 cdp 6 sx 3262 gx 100 offset 3162 ep 102 cdp 6 sx 3237 gx 125 offset 3112 ep 101 cdp 6 epi sx 3637 gx 825 offset 2812 ep 117 cdp 50 sx 3612 gx 850 offset 2762 ep 116 cdp 50 Sx 3587 gx 875 offset 2712 ep 115 cdp 50 Sx 3562 gx 900 offset 2662 ep 114 cdp 50 Sx 3537 gx 925 offset 2612 ep 113 cdp 50 sx 3512 gx 950 offset 2562 ep 112 cdp 50 sx 3487 gx 975 offset 2512 ep 111 cdp 50 sx 3462 gx 1000 offset 2462 ep 110 cdp
169. nvestigators in oil companies implemented variations of Hagedoorn s graphical migra tion using notions from the theory of wave propagation and methods from signal pro cessing theory May different types of migration were thus created By accident some of these methods were amplitude preserving meaning that reflectivity information is preserved in the image produced by such a migration Such true amplitude or amplitude preserving migrations became important when issues of reservoir characterization by seismic methods became important The first of these reservoir characterization methods first discovered in the 1970s was called the bright spot method which allowed the identification gas sands in the Gulf of Mexico by their high amplitude reactivities In reality all that was done to see the bright spots was for seismic data processors to stop normalizing away the amplitude differences in their migrated images This change marked the beginning of seismic migration as a parameter estimation tool 103 Chapter 8 Zero offset v t and v x z migration of real data Lab Activity 9 Now that you have had an introduction to a few zero offset migration codes we will apply some of these migrations to a dataset consisting of real data As before make a temporary directory in a scratch directory area Call this one Temp3 cd scratch yourusername mkdir Temp3 cd Temp3 cp data cwpscratch Data3 Stoltmig cp data
170. o 1500 1800 2300 tnmo 0 0 1 0 2 0 Figure 9 7 CV stacks and a Brute stack a Raw stack no NMO correction b vnmo 1500 c vamo 2300 d vamo 1500 1800 2300 tnmo 0 0 1 0 3 0 130 9 7 Quality control through raw CV and brute stacks The term quality control or QC is the industry name for what we have been calling getting to know your data The most widely used method of QC is to perform a CV or brute stack of the data 9 7 1 Lab Activity 15 Raw Stacks CV Stacks and Brute Stacks Another common quick look technique is the construction of brute stacks As the name suggests a brute stack is a stack of CMP data without a correct NMO correction Typically some form of brute stack is used as a field quality control technique A raw stack For example we may use sustack to stack the CMP gathers with no NMO correction Because industry uses the term brute stack with the assumption that some rough NMO correction is applied we use the term raw stack for the case of a no NMO stack of the data For example try sustack lt gain jon 1 cdp su suximage perc 99 title Raw Stack amp This will take a few minutes to come up on the screen Remarkably we can see the hint of our structure even in a stack with no NMO correction So yes we have data Constant Velocity stacks guessed stacking velocities We can answer other questions for example we might want to know more about the multiples We can NMO cor
171. o say the process of removing the effect of a waveform to produce a desired output Symbolically we have recorded data D t with a particular waveform W t possibly distorting the arrival time of a given reflection What we want ideally is to reconstruct the reflectivity series by applying the inverse process S t W t x D t We need only define what the inverse of W t given by W 1 t is If we write this out in the Fourier domain representation then we see that S t Wx D t T W1 t1 D t ti dt 1 fe n s w edw In Joo w w Thus we see that deconvolution is division in the frequency domain 169 Deconvolution of functions represented by their Z transforms In terms of Z transforms we would then be dividing by the polynomial representation of the signal The zeros of the Z transform polynomial become poles in the deconvolution result which if were were performing the inversion by contour integration would be the contribution to the contour integral 11 2 4 Division in the frequency domain There is a problem however when we consider the Fourier transform as a spectrum If the function is zero over a range of values in the frequency domain as opposed to isolated zeros in the Z transform in the Fourier transform form of the wavelet given by the function w w then deconvolution is unstable or undefined This follows because division by a small number introduces computational instability and of course
172. o see the meanings of the graphical convention used on a plot Try suximage lt sonar su legend 1 amp This will show a grayscale bar There are a number of colorscales available Place the mouse cursor on the plot and press h you will see that further pressings of h will re plot the data in a different colorscale Now press r a few times The h scales are scales in hue and the r scales are in red green blue rgb It is important to see that the brightest part of each scale is chosen to emphasize a different amplitude With colormapping some parts of the plot may be emphasized at the expense of other parts The issue of colormaps often is one of selecting the location of the bright part of the colorbar versus darker colors Even perfectly processed data may be rendered uninterpretable by a poor selection of colormapping This effect may be seen in Figure 3 4 Repeat the previous this time clipping by percentile suximage lt sonar su legend 1 perc 99 amp The ease at which colorscales are defined and the fact that there are no real standards on colorscales mean that effectively every color plot you encounter requires a colorscale for you to be able to know what the values mean Furthermore some colors ranges are brighter than others By moving the bright color to a different part of the amplitude range you can totally change the image This is a source of richness of display but it is al
173. oD H 143 velocity m s 2500 3000 a velocity m s b 5 500 2000 2500 3000 500 2000 time s wo time s wo Semblance plot no multiples Semblance plot data w b multiples C velocity m s 500 2000 2500 3000 3500 time s wo Semb plot data w b pegleg multiples Figure 10 6 a Synthetic data similar to CDP 265 b Synthetic data plus simulated water bottom multiples c Synthetic data plus water bottom multiples plus select pegleg multiples 144 1000 0 1000 a p intercept time in ms b p intercept time in ms 2000 1000 0 1000 2000 tau s tau s Synthetic data in the tau p domain Synthetic data w b multiples tau p domain p intercept time in ms C 7000 1000 0 1000 tau s Synthetic data w b pegleg multiples tau p domz Figure 10 7 a Synthetic data in the Radon domain b Synthetic data plus simulated water bottom multiples in the Radon domain c Synthetic data plus water bottom multiples plus select pegleg multiples in the Radon domain 145 suximage d2 15 2 1450 verbose 1 title faketwater cmap hsv2 legend 1 bclip 5 amp suvelan nv 150 fv 1450 dv 15 lt faketwatertpegleg su suximage d2 15 2 1450 verbose 1 title faketwater tpegleg cmap hsv2 legend 1 bclip 5 amp Plots like these are shown in Figure 10 6a b c In Figure 10 6 a we see the velocity analysis semblance plot for the synthetic data without multiples This wo
174. ocity change across the profile The Velan script runs a number of SU programs to make an estimate of the velocity profile in time and when you are finished doing velocity analysis will generate uniformly sampled velocity profiles of both RMS and interval velocities These are quick and dirty representations of the velocity field which tend to be bad because they have spurious errors due to errors in interpo lation and in conversion from RMS to interval velocities These can be used as a guide for the construction of a background velocity model This procedure may be repeated changing for example the value of the cdp incre ment dcdp 256 dcdp 128 or dcdp 64 to obtain increasingly dense coverage in RMS velocities The idea is to go in increments of powers of 2 so that you only have to do every other CDP Or the user may start with a different value than 512 and bisect that The Velan script will produce a collection of average and interpolated RMS and interval velocities calculated using the programs unisam and velconv These velocity files will contain errors owing to errors in the interpolation and probably should be used with caution for further processing but do provide a starting point We may view the resulting velocity files by running the shell script Xvelocity which will show RMS average RMS interpolated RMS as well as average interval velocities These automatically generated velocity files tend to be lumpy so they are not
175. of d ld yd and dj t o arctan d We can experiment with spectral whitening in SU using the command suwfft This program gives the user the choice of whitening from moderate to extreme The plots labeled traditional use the default settings of the program which are not really full spectrum whitening For example suwfft lt fake su w0 0 wi 1 w2 0 suifft suxwigb xcur 2 title fully white spectrum suwfft lt fake su suifft suxwigb xcur 2 title traditional spectral whithening and 165 suwfft lt gain jon 1 cdp 265 su w0 0 wi 1 w2 0 suifft suxwigb xcur 2 title fully white spectrum suwfft lt gain jon 1 cdp 265 su w0 0 wi 1 w2 0 suifft suxwigb xcur 2 title traditional spectral whitening show us what this does to our data The data are sharpened but then so is the noise Another test is to apply spectral whitening to our CMP 265 data Recall that fre quency filtering may need to be applied before and after the whitening process We see that the spectrum is an idealized white amplitude spectrum whose shape is the filter sufilter f 0 5 80 90 lt gain jon 1 cdp 265 su suwfft w0 0 wi 1 w2 0 suifft sufilter f 0 5 80 90 suspecfx suxgraph title Spectrum after whitening amp sufilter f 0 5 80 90 lt gain jon 1 cdp 265 su suwfft suifft sufilter f 0 5 80 90 suspecfx suxgraph title Spectrum after whitening amp and though it appears to be a
176. ore OO0ar Gok Dy 250d tl a a Erk GOAN Re Se 189 12 2 2 Applying migration 4 4 00 4a we Be ah a ea A ss Gee 189 12 3 Homework Assignment 10 Preliminary Stolt migration Due 1 Nov 2012 pefore 92 QU arm s ra ce se le an tet Bins es edwin owes os Abn he 28S od 190 T24 Other velocity iles i mata ma 44 ee eh SNe oe Bei By ee AR ke hd 191 12 4 1 Velocity analysis with constant velocity CV stacks 191 12 5 Dip Moveout DMO 4 2 2 8 leon O88 bok ag go Eo ese ee Ae 193 12 5 1 Implementing DMO lt x ieee ee ee ee Ee eek Bee 193 12 6 Concluding Remarks 262 2654 RS See ea Ree Bae es 194 13 Velocity models horizon picking and muting 195 T31 Honzon picking asg a aip es ape eek Redes foe Pete ep a eee 196 13 2 Migration velocity tests 2 Gag aa a oS ease hele ee 8 197 13 2 1 Homework 11 Build a velocity model and perform Gaussian Beam Migration Due 15 Nov 2012 198 13 3 Concluding remarks se oaoa aa a Ae ape ae a tgs ad 198 14 Prestack Migration 199 14 1 Prestack Stolt migration 44 5 5 Nod ahd ek Beer a ee ne es 199 14 2 Prestack Depth Migration 2 2 4 61 2 teste 4b oe 4G ae 4b ae hd 200 14 3 Concluding remarks iy ce oe gt aoe e 4 gl andy eee ch a eer eek 200 List of Figures 1 1 2 1 2 2 2 3 2 4 3 1 3 2 3 3 3 4 3 5 3 6 3 7 5 1 5 2 A quick reference for the vi editor 000 0000 5a The suplane test pattern aaa aaa a a The suplane test pattern b the Fourier
177. orking shell looks for executable files in Usually executables are stored in a sub directory bin of some directory Because there may be many software packages installed on a system there may be many such locations To find out what paths you can access which is to say which executables your shell can see type echo path or echo PATH 6699 The result will be a listing separated by colons of paths or by spaces to executable programs 1 10 2 The CWPROOT variable The variable PATH is important but SHELL and PATH are not the only possible envi ronment variable Often programmers will use an environment variable to give a users shell access to some attribute or information regarding a specific piece of software This is done because sometimes software packages are of restricted interest For SU the path CWPROOT is necessary for running the SU suite of programs We need to set this environment variable and to put the suite of Seismic Unix programs on the users path 1 11 Shell configuration files Because the users shell has as an attribute a natural programming language many configurations of the shell environment are possible To find the configuration files for your operating system type ls a lt show directory listing of all y 8 files and sub directories pwd lt print working directory then the user will see a number of files whose names begin with a dot 20
178. ow for suxwigb but everything else should be ok To get your email you may need to have a file called forward dot forward in your home directory on the lab machines with your email address e g youruser name mymail mines edu as the single line of text in the file Killing an at job We all make mistakes Sometimes we launch at that we want to remove If it turned out that you changed your mind and didn t really want to run the job you would type atq 2 Wed Oct 7 01 00 00 2009 a john atrm 2 where this is the job id number that is in the first line of the atq output 10 5 Concluding remarks Our discussion of velocity analysis and multiple suppression by the radon transform here is just a warm up for a more production level treatment in Chapter 12 Obviously there are many issues that come into play when performing velocity anal ysis We do velocity analysis for multiple suppression but we also are doing velocity analysis for stacking We do not know a priori what the correct NMO velocities are so there is an iterative aspect to the process Once we have suppressed the multiples we may recognize that our gaining is not very good and we likely need to regain the data and pick NMO velocities again Remember also that we should have muted certain arrivals at the beginning of all of these operations Finally there is the issue of time Time is money We cannot afford to be perfection ists There is a line that has to b
179. pacing sufficiently fine to prevent spatial aliasing However there are situations particularly in the case of out of plane noise events that noise can be spatially aliased Furthermore we may have missing traces that cause artifacts in migration 7 11 3 Recognizing spatial aliasing in the f k domain The term spatial aliasing implies that there is a spatial subsampling which occurs imply ing that there is a wrapping of data in the wavenumber domain We can see an example of this in the comparison of f k transforms of simple su data with the interp su data in Figure 7 2 In Figure 7 2a and b the simple su and interp su data are shown in the f k domain The interp su data show a typical f k domain representation of seismic data As with aliasing in the frequency domain we see that as we reach the maximum wavenumber the equivalent of the Nyquist frequency in the w k domain the spectrum is still nonzero Typically we like to have data vanish smoothly at the Nyquist frequency or wavenumber to ensure the absence of aliasing without introducing ringing in the data Meaning of the wavenumber domain But what does the wavenumber domain mean As we will see in the next section there is a relationship between the magnitude of the k vector and the quantity w c in data that are governed by a wave equation For zero offset reflection seismic data we actually have k 2 w c were the factor of 2 comes from the fact that we are dea
180. perform many configuration changes and personalizations to your working environment which can enhance your user experience For these notes we concentrate only on enough of these to allow you to work effectively on the examples in the text 1 7 Choice of editor To edit files on a Unix like system the user must adopt an editor The traditional Unix editor is vi or one of its non proprietary clones vim vi improved gvim or elvis The 14 vi environment has a steep learning curve making it often unpopular among beginners If a person is envisioning working on Unix like systems a lot then taking the time to learn vi is also time well spent The vi editor is the only editor that is guaranteed to be on all Unix like systems All other editors are third party items that may have to be added on some systems sometimes with difficulty Similarly there is an editor called emacs that is popular among many users largely because it is possible to write programs in the LISP language and implement these within emacs There is a steep learning curve for this language as well There is often substantial configuration required to get emacs working in the way the user desires A third editor is called pico which comes with a mailer called pine Pico is easy to learn to use fully menued and runs in a terminal window The fourth class of editor consists of the screen editors Popular screen editors include xedit nedit and gedit There is a win
181. pershot by putting some display gain on this with sugain and display with suxwigb 154 wide angle reflections merencs batt wy mute above this curve Supershot Figure 10 10 An average over all of the shots showing direct arrivals head waves wide angle reflections and a curve along with muting may be applied to eliminate these waves sugain jon 1 lt supershot su suxwigb key offset It is important for picking that suxwigb is run with key offset so that the horizontal scale will be in offset 10 3 5 How to pick mute values On the super shot gather we may pick time offset pairs to supply to sumute This may be done by placing the cursor on the desired point and typing the letter s The ordered pair of time and offset will be printed on your terminal window Once the desired times and offsets are obtained the entire dataset may be muted via sumute lt seismic su tmute t1 t2 xmute x1 x2 key offset gt mute seismic su 155 10 3 6 The shape of the wavelet It may have occurred to the reader that we have done nothing to ensure that the waveform is actually optimal for stacking Seismic signals are assumed to have what is called the minimum phase property which is to say that most of the energy is located at the beginning of the wave form Manufacturers of marine air guns do their best to fit this prescription but there are still issues that make real seismic data
182. pth migration For example we may perform Stolt migration and convert this to a depth section with suttoz to obtain first approximation of the migrated image sustolt par stolt par cdpmin 1 cdpmax 2142 dxcdp 12 5 lt stack nmo radon gain jon 1 su gt stolt seis su suttoz par suttoz par nz 1500 lt stolt seis su gt stolt depth seis su The file suttoz par may be made by running shell script RMStoINT which employs suintvel to do the interval velocity conversion This script is provided in data cwpscratch Data5 The shell script produces as its output the file suttoz par The contents of this simple shell script are bin sh convert a single RMS velocity to Interval velocity put your values of tnmo and vnmo here tnmo vnmo output file outfile suttoz par suintvel t0 tnmo vs vnmo outpar outfile echo t tnmo gt gt outfile 195 echo use v and t values for suttoz echo result output in outfile exit 0 The entries for vnmo and tnmo are the same values that you used for vmig and tmig in stolt par 13 1 Horizon picking From the contour map of vintzx bin that we view with Xvelocity one of the other velocity plots that is output by Velan or from independent well log information we may be able to get an idea of a reasonable velocity trend that will be appropriate for migration If you know the stratigraphy of an area reasonable velocities estimates for given rock types provide this type of
183. pthref 1000 pmin 2000 pmax 2000 dp 8 igopt 2 we rely on the choice of tnmo and vnmo to separate data and multiples pmula 20 pmulb 200 filter data in radon domain choose 1 data multiples sunmo par parfile lt data suradon offref offref depthref depthref pmula pmula pmulb pmulb interoff interoff pmin pmin pmax pmax dp dp choose choose igopt igopt sunmo par parfile invert 1 gt radondata 187 clean up bin rm f radontmp exit 0 Before proceeding further we may want to manually edit the radonnmovel par or numovel par so that the values make sense For example in an ocean survey such as our Viking Graben dataset we always have a velocity of 1500 m s at time t 0 0 12 2 NMO and Stack As before we can perform the NMO followed by Stack of our data via sunmo par nmovel par lt radon gain jon 1 cdp su sustack gt stack nmo radon gain jon 1 su The file nmovel par has contents that are similar to cdp 128 192 256 320 384 448 512 576 640 704 768 832 896 960 1024 1088 tnmo 0 0 0 636979 1 15987 1 31199 1 91094 2 9187 vnmo 1500 1557 09 1734 03 1822 5 2041 34 2488 34 tnmo 0 0 0 579936 0 893672 1 35002 1 8539 2 41482 2 7951 vnmo 1500 1561 75 1659 53 1831 81 2032 03 2446 44 2693 22 tnmo 0 0 0 598951 0 922194 1 35952 1 50213 1 89192 2 90919 vnmo 1500 1557 09 1678 16 1794 56 1855 09 2027 38 2562 84 tnmo 0 0 0 684515 1 29297 1 87291 2 3958 vnmo 1496 56 1571 06 1734
184. put wavelet in SU SEGY trace format d vector of desired output wavelet or be ONE a Sx dfile file containing desired output wavelet in SU format dt tr dt if tr dt is not set in header then dt is mandatory Optional parameters nshape trace length of shaping filter pnoise 0 001 relative additive noise level showshaper 0 1 to show shaping filter For example our waveforms may be shaped via sushape dfile dfile su wfile wavelet su lt data su gt shaped_data su The shaping filter works by effectively by performing the operation of deconvolving the data to remove wavelet su and the convolution of the resulting spiked data by the desired output waveform dfile su The sushape program makes use of Wiener Levinson theory to perform this operation in the time domain Finding the wavelet and making target waveforms A great deal of work has been put into wavelet estimation techniques in the explo ration seismic community Ideally we should know the wavelet for each shot and even the wavelet as a function of angle from the shot Here we assume for simplicity that waveforms chosen carefully off of the data using suwind are sufficient for our purposes To construct the target waveform dfile su we may use one of the wavelets generated from the program suwaveform with either type ricker1 or type ricker2 being the best choices although the Berlage waveform is not a bad choice either When con structing a target wavefor
185. r systematic errors in the data 9 4 1 Viewing a specific Shot gather Close the movie window and capture a particular shot gather for study For example we will capture the shot gather at ep 200 but any will do This is done via suwind 116 a offset meters b trace number 6 3000 2000 1000 20 40 60 80 100 shot ep 200 shot ep 200 Figure 9 2 a Shot 200 as wiggle traces b as an image plot suwind lt seismic su key ep min 200 max 200 gt shot ep 200 su This will take a few minutes Once you have this shot gather you may view the data both as an image plot and as shot gather suximage lt shot ep 200 su perc 99 title shot ep 200 amp suxwigb lt shot ep 200 su perc 99 title shot ep 200 amp The view of the data is not particularly good because we have not applied a gain to the data to take into account the amplitude decay with distance traveled 9 4 2 Charting source and receiver positions We may view a chart of the source receiver positions with suchart This is done via suchart lt seismic su xgraph n 120120 linewidth 0 label1 sx label2 gx ma
186. r this example You may need to experiment with this parameter on your data The entries in dmofk par must be named tdmo and vdmo for sudmofk to be able to see them The DMO process may take some time to complete time to complete After the process has finished you will need to resort your data back into CMP gathers via susort cdp offset lt dmofk nmo radon gain jon 1 co su gt dmofk nmo radon gain jon 1 cdp su 193 These new data may then be stacked and migrated Again you must have sufficient storage capacity to save a couple of copies of the full dataset We expect improvement of the stack with DMO only in areas where dips are suffi ciently large that the NMO approximation fails to completely flatten the data If the input data are reflections over a generally flat or low dip geology then we don t expect to gain much by performing this operation 12 6 Concluding Remarks The notion of NMO followed by DMO as a way of building equivalent zero offset datasets naturally let to the notion of transformation to zero offset TZO and migration to zero offset MZO The basic notion of TZO and MZO is the following Suppose that you could do a perfect amplitude preserving prestack depth migration of data with offset Then the result would be a representation of the earth a geologic model with bandlimited delta functions describing the reflectors Then suppose we did a remodeling from that representation to yield zero offset
187. rable to work in these directories and save only really important materials in your home directory Users should be aware that administration of scratch directories may not be user friendly Using up all of the space on a partition may have dire consequences in that the 18 administrator may simply remove items that are too big or have a policy of removing items that have not been accessed over a certain period of time A system administra tor may also set up an automated grim file reaper to automatically delete materials that have not been accessed after a period of time Because files are not always automatically backed up and because hardware failures are possible on any system it is a good idea for the user to purchase USB storage media and get in the habit of making personal backups on a regular basis A less hostile mode of management is to institute quotas to prevent single users from hogging the available scratch space You may see a scratch directory on any of the machines in your lab but these are different directories each located on a different hard drive This can lead to confusion as a user may copy stuff into a scratch area on one day and then work on a different computer on a different day thinking that their stuff has been removed The availability and use of scratch directories is important because each user has a quota that limits the amount of space that he or she may use in his her home directory On systems wh
188. rated version of your data from Homework 10 into a depth section e Use this depth section as input to the Horizon script and build a velocity model by picking several horizons in the data Assign interval velocities based on the average interval velocities that are shown when you run Xvelocity e Run the Unif2 sh script to generate the smoothed velocity files unewvelxz bin and unewvelzx bin e Use the unewvelzx bin file and your stacked data in the script Gbmig to perform Gaussian beam migration on these data 13 3 Concluding remarks Seismic data and any other auxilliary data are combined in a step that involves the intelli gence of the processor to build better background velocity profiles for further migrations and for modeling We should not expect that this process can be made totally automatic Indeed some experimentation will show that it is quite easy for iterations of picking and migrating to yield a diverging series of modifications that would ultimately result in a terrible and unrealistic velocity profile In the modern world new techniques such as full waveform inversion provide an ap proach that yields estimates of velocities The preferred method of migration is reverse time migration RTM which is a prestack depth migration method 198 Chapter 14 Prestack Migration Owing to advances in computer technology it is now possible to perform prestack migra tion The motivation for doing this is that information rega
189. rd of the down traveling field All we have is the record of that part of the reflected field that hits the surface of the earth where there are geophones The migration process finds the place where the downward travelling field and the reflected field overlay the reflector surface One way of looking at migration is that we would like to cross correlate the down traveling field with the time reversed reflected field The place where these fields correlate is at the reflector surface You may also see what the seismic data looks like recorded at the surface of the earth model by viewing the file hseis su via suximage lt hseis su perc 99 to see the direct arrival and the bowtie shaped reflected arrival Finally you may change the depths in the model by editing the file syncline unif2 or change the location of the source to see what varying these quantities changes in the data You may slow down the movie by increasing the value of the sleep parameter 6 2 Lab Activity 5 Hagedoorn s graphical migration The purpose of this lab example is to migrate the simple data in Figure 6 1a by Ha gadoorn s graphical method These synthetic data represent the zero offset reflection seismograms recorded over the undulating reflector model in Figure 6 1b The wave speed in the upper medium is assumed to be 2000 m s and the data are drawn in such a way that 1 0 s two way time is equivalent to 1000 m of distance Thus the time scale translates
190. rding the angular dependence of the reflectivity is preserved in prestack methods but is not preserved in poststack mi gration Having such information can give clues about the material properties at depth that are helpful in determining porosity and permeability as an aid in reservior charac terization Furthermore prestack migration may be used as a step in migration velocity analysis If the output is in the form of image gathers which is to say migrated CMP gathers then analysis of the curvature of the arrivals in the image gathers can be used as a guide for updating the wavespeed profile Finally prestack migration is preferable if there is large vertical relief in subsurface structures Data which are largely flat do not benefit from prestack migration The cost of prestack migrating a single CMP gather is comparable to migrating the entire poststack profile so the computer costs increase correspondingly 14 1 Prestack Stolt migration The program sustolt may be used to perform prestack time migration by first sorting the multiple suppressed gained and NMO corrected data into common offset gathers and then applying sustolt par stolt par cdpmin 1 cdpmax 2142 dxcdp 12 5 lt nmo radon gain jon 1 co su gt prestack stolt nmo radon gain jon 1 su This takes about an hour to complete on a relatively fast PC The result can be sorted into image gathers which is to say migrated CMP gathers The motivation to do such a thing may
191. rect the data to the water speed of 1500 m s to view the multiples type sunmo vnmo 1500 lt gain jon 1 cdp su sustack suximage perc 99 title CV stack vnmo 1500 amp We look for repetition in the data Multiples consist not only of bounces from the first arrival in the water column but multiple bounces of many arrivals The water surface is a big mirror and not only are there multiple reverberations of the first arrival but multiple reverberations of most arrivals such that potentially the whole seismic section is repeated with multiples of early arrivals overwriting the later arrivals If we choose a number that corresponds to the moveout time for later arrivals for ex ample vnmo 2300 this will will tend to enhance later reflections though it is apparent that reverberations dominate the image sunmo vnmo 2300 lt gain jon 1 cdp su_ sustack suximage perc 99 title CV stack vnmo 2300 amp 131 Putting it together a brute stack Suppose that we guess a profile with NMO corrected using vnmo 1500 1800 2300 set at the at the times tnmo 0 0 1 0 2 0 sunmo vnmo 1500 1800 2300 tnmo 0 0 1 0 2 0 lt gain jon 1 cdp su sustack suximage perc 99 title vnmo 1500 1800 2300 amp This choice of velocities focuses both the water bottom and shallow section as well as the strong reflector that starts at 2 0 seconds on the left of the section But note these values are really just guesses Even with a more realistic veloc
192. relatively strong reflectors Eventually the velocity will tend to increase slowly Look in Oz Yilmaz book or in Hill and Rueger s notes for an example of perfect semblance plots What if it doesn t work The only thing left to vary is the location of the filter in the Radon domain The filter is defined by the values of pmula and pmulb in as used in suradon The filter is defined by a straight line with endponts time 0 and the value of pmulb and time maximum time on the section and the value of pmula In practice we have found that good values of pmulb are between 180 and 240 and good values of pmula are between 5 and 30 Feel free to experiment Note that the semblance peaks move to the right or the left a bit depending on where in the radon domain the filter is applied So you really need to repick your velocities after you have changed the filtering 151 10 2 2 Are we finished with multiple suppression and velocity analysis You will notice a that on the near offsets there are just as many multiples as when we started This occurs because at near offsets all arrivals are almost flat whether they are once reflected arrives from beds or if they are multiples Differential moveout methods are less effective at near offsets 10 3 Muting revisited If you plot the multiple suppressed gather from the Homework assignment you will notice that values have been zeroed for short times and far offsets Thus there are some
193. rength and receiver gain Again we are free remove the separate shot files after the process is complete via rm split_ Both of these operations are captured in the shell script Pbal located in data cwpscratch Data5 11 8 Muting NMO corrected data The program sumute may be used to surgically remove undesirable noise on the CMP gathers that occurs for times early than the water bottom reflection Because our prospect has a roughly flat surface the time of the reflection of the water bottom is at approximately time 48 seconds In addition to the noise before the water bottom reflection there are some unsuppressed multiples or other arrivals on the far offsets that are undesirable We may use predictive deconvolution to clean up those near offset traces or we may consider eliminating the near offset traces entirely with suwind before further processing Putting these together after NMO we may insert the commands suwind key offset min 3237 max 450 sumute key offset tmute 45 45 xmute 3237 450 prior to the stack to clean up the image The choice of nearest offset to include is a matter of personal preference The value of 450 is not necessarily the best value 11 9 Surface related multiple elimination A modern approach to multiple elimination is the Surface Related Multiple Elimination SRME method invented in 1991 by Erich Verschuur then a Ph d student at Delft University of Technology The method is a data driven
194. res simply knowing the speed of sound in water and the speed of light respectively This sounds simple and if we had all of the information in a neat and consistent form it would be The first complication comes from the fact that SU is a seismic package When non seismic data are used in a seismic package often the time sampling interval must be scaled to store the data The reason for this is to save storage space in the header the creators of the SEG Y data format chose the time sampling interval dt not to be a floating point number but rather as an unsigned short integer On a 32 bit machine the size of the largest value that an unsigned short can take on is 65535 Thus scaling is necessary Usually these scale factors are multiples of some power of 10 Try doing depth conversion on the sonar and radar data using values you know for the speed of sound suttoz v SPEED_OF_SOUND_IN_WATER lt sonar su suximage perc 99 amp and the speed of light suttoz v SPEED_OF_LIGHT lt radar su suximage perc 99 amp respectively The speed of light is 2 998 x 10 m s The speed of sound in water is 1500m s Likely the correct values to use in each case will be off by some multiplier that is a power of 10 owing to the fact that the natural frequencies available for radar and sonar are not in the same band as those used for seismic data If we type sukeyword dt unsigned short dt sample interval in micro seconds we see th
195. rforming this procedure in subsets makes it a bit quicker this is for instruc tional purposes only From now on we work with the full dataset We do not break the data into blocks 183 11 11 Concluding Remarks Much of exploration seismic reserarch conducted prior to the mid 1980s was focused on the problem of seismic deconvolution and wavelet estimation The CWP SU Seismic Unix package was largely developed during a time right after this when exploration seismic research was focused on amplitude preserving depth migration so consequently we have a lot of migration related tools and a comparatively few deconvolution related programs so deconvolutional methods are not yet well represented in the SU package Though we have not really done justice here to the broad topic of deconvolution and the other spectral methods we can see that the application of these techniques is more involved than merely applying the operation Considerable preconditioning of the data is required to make the traces look more alike so that the deconvolutional process may remove the parts such as multiples that we don t want Predictive decon really means that we use the first part of the data to predict the repetitions in the latter part of the data and to use those predictions to annihilate those repetitions multiples For this to work the repetitions must closely match the initial waveforms Hence making the amplitudes as uniform as possible is desireable for
196. rksize 2 mark 8 title sx gx chart amp If you zoom in on the plot missing shots are revealed Another popular type of chart called the stacking chart is discussed below The suchart program is useful as a quality control tool because any errors in the headers or inconsistencies in the data are immediately revealed by the plot of header values 117 9 5 Geometrical spreading aka divergence correction The amplitudes of seismic waves experience a reduction in amplitude that is a function of the distance r that the wave travels There are two sources of this amplitude reduction The first is due to geometrical spreading The wave energy remains constant but as the wavefront expands the energy density reduces as a function of the increasing area of the wavefront 9 5 1 Some theory of seismic amplitudes For constant velocity solutions to the constant velocity scalar wave equation 1 0 v Ti U x y z t o Yo 20 W t d x xo ly yo z 2 look like 1 WEAHS U x Z t Xo 2 x y 0 Yo 20 Inr where x y z is the coordinates of a point on the wavefront o Yo Zo are the coordi nates of the source t is traveltime c is the constant wavespeed we assume that the source starts at time t 0 and r y x z0 y yo z 20 is the radial distance from the source point to a point on the wavefront W t r c is a waveform traveling at speed c which arrives at tim
197. rograms to lower case preserving strings FCAT fast cat with 1 read per file ISATTY pass on return from isatty 2 MAXINTS Compute maximum and minimum sizes for integer types PAUSE prompt and wait for user signal to continue T time and date for non military types UPFORT change Fortran programs to upper case preserving strings In CWPROOT src par main A2B convert ascii floats to binary B2A convert binary floats to ascii CSHOTPLOT convert CSHOT data to files for CWP graphers DZDV determine depth derivative with respect to the velocity FARITH File ARITHmetic perform simple arithmetic with binary files FINSTRIP convert a file of binary data plus record delimiters created FINUNSTRIP convert C binary floats to Fortran style floats GRM Generalized Reciprocal refraction analysis for a single layer H2B convert 8 bit hexidecimal floats to binary More 3 Hitting the space bar shows the rest of the page The suname output shows every library function shell script and main program in the package and may be too much information for everyday usage What is more common is that we might want a bit more information than a selfdoc but not a complete listing This is where the sudoc feature is useful Typing sudoc NAME yields the sudoc entry of the program NAME For example we might be interested in seeing information about suplane sudoc suplane and comparing that with
198. s These reflectors are the boundaries between media of different wavespeeds Given auxiliary information about geology well logs or the result of seismic velocity analysis we expect to be able to relate arrivals on the seismic section to specific depth horizons for which in turn we have wavespeed information 5 4 Time to depth conversion of a test pattern To see what the problem of time to depth and depth to time is all about we may try suttoz on a test pattern made with suplane suplane gt junk su suttoz lt junk su t 0 0 15 2 v 1500 2000 3000 gt junki su suxwigb lt junk su title test pattern amp suxwigb lt junki su title depth section amp The program suztot has been provided to apply depth to time conversion as the inverse of sutotz Because we know the values of the velocity that were used we must 61 trace number trace number a 10 20 30 b 10 20 30 0 0 gt 0 054 1004 by h f i by 7 0 104 ans I 2 aadi 244143 5 o gt E r amp 200 0 154 as t gt 0 20 300 0 254 test pattern depth section c trace number 10 20 30 ae time s ese ea ed om a a e a
199. s how much of the output to show You may try different values of dz and nz to see what happens to your depth stretched image 12 4 Other velocity files There are several velocity files that are generated by the Velan script These velocity files are provided only as a guide for later velocity model building The file vintzx bin is a very approximate Vint z but likely contains irregularities that would not make this the best velocity file to use for migration though it is in the correct format to be used as the input vfile in sumiggbzo or for the input velocity file for rayt2d which generates the traveltime tables for sukdmig2d Similarly while the file vintxz bin is in the correct format to be used as the input for sumigfd sumigffd sumigsplit or sumigpspi it too will likely be too lumpy to give a good result We can however use these estimated velocities for some interval velocity information when we build velocity models by horizon picking 12 4 1 Velocity analysis with constant velocity CV stacks We have used a method of semblance picking to estimate the stacking velocities This is an estimate but it is not the only way to get the stacking velocities An alternate method is the constant velocity stack CVS In data cwpscratch Data5 are two shell scripts MakeC VSstackMovie bin sh data multiple_suppressed_gained_data su movie stackmovie su fvel 1500 dvel 10 vmax 3000 vel fvel rm movie 191
200. section Note that the curvature seen depth section indicates a non piecewise constant u t Note that the reconstructed time section has waveforms that are distorted by repeated sinc interpolation The sinc interpolation applied in the depth to time calculation has not had an anti alias filter applied 2 2 35 37 38 39 Al 42 60 5 3 6 1 6 2 6 3 6 4 6 5 6 6 6 7 6 8 6 9 a Cartoon showing an idealized well log b Plot of a real well log A real well log is not well represented by piecewise constant layers c The third plot is a linearly interpolated velocity profile following the example in the text This approximation is a better first order approximation of a real sell logs stiess 2 AS 4 oh Sk eek e e a ee Sig ee EEs a Synthetic Zero offset data b Simple earth model The Hagedoorn method applied to the arrivals on a single seismic trace Hagedoorn s method applied to the simple data of Fig 6 1 Here circles each centered at time t 0 on a specific trace pass through the maximum amplitudes on each arrival on each trace The circle represents the locus of possible reflection points in x z where the signal in time could have o einatedi 12 eo ee eg ees Oe Be ol te ee The dashed line is the interpreted reflector taken to be the envelope of the CiiGles y es 2a te AIS bee ek Ghee a Melt Dale mesa The light cone representation of the constant velocity solution of the 2D wave equat
201. sentation of a function is a frequency domain representation as a polynomial the division by the polynomial representation of the waveform is deconvolution To be minimum phase the poles and the zeros of the inverse function must also be inside the unit circle If we think of a simple example of what this means consider the function Z As a signal this would be 0 0 1 0 0 which is a unit spike that appears at sample position n A unit spike is interpreted as being a Dirac delta function and is thus the most compact waveform possible With it s energy concentrated at a point it fits the description of a minimum phase waveform 162 a amplitude b amplitude 1000 2000 3000 1000 2000 3000 o S 5 60 O o 80 100 100 120 120 spectra spectra Figure 11 1 a Amplitude spectra of the traces in CMP 265 b Amplitude spectra after filtering The function Z has n zeros at Z 0 Deconvolution by Z is division by Z The resulting filter 1 Z now is seen to have an n th order pole at Z 0 which again fits our minimum phase description of poles being inside the unit circle Because Z exp iw the value of Z never actually passes through a zero when considered in terms of the w domain 11 1 3 White spectrum Because deconvolution may be thought of as division in the frequency domain we really cannot have the function we are dividing by be zero in any place that will affect the o
202. single curve these are really 55 identical spectra on on top of the other for each of the 55 traces in our gather The spectral whitening process involves the cascade of forward and inverse Fourier transforms Owing to zero padding in these transforms there may be more samples per trace on the output than on the input so an extra step of windowing the data with suwind in time is required sufilter f 0 5 80 90 lt gain jon 1 cdp 265 su suwfft suifft sufilter f 0 5 80 90 suwind itmin 1 itmax 1500 suxwigb title data after traditional spectral whitening amp The windowing passes samples from sample 1 through sample 1500 on each trace As with our fake data the real data have additional arrivals We can run Radon test on versions of the data after spectral whitening has been applied Observe the changes in the semblance plots after spectral whitening It may be that we would prefer to use spectral whitening after multiple suppression 11 1 6 Spiking deconvolution also known as minimum phase deconvolution or prediction error filtering We have another way of whitening the spectrum This method is to deconvolve the data by a statistical estimate of the wavelet which is based on the assumption that the data are minimum phase Thus we assume that our entire data consists of spikes convolved with minimum phase wavelets Before launching into the application of minimum phase deconvolution we discuss the operations that we will be
203. so a potential source of trouble if the proper balance of color is not chosen 3 4 Normalization Median balancing A common data amplitude balancing is to balance the colorscale on the median values in the data The median is the middle value meaning that half the values are larger than the median value and half the data are less than the median value Thus the traces are normalized by this middle value Another possibility is to scale traces by dividing by some constant value For example dividing each trace by the square root of the average of the sum of the square of its values RMS Type these commands to see that in SU sunormalize norm med lt sonar su suximage legend 1 sunormalize norm med lt sonar su suximage legend 1 perc 99 36 400 200 99 and legend 1 Figure 3 3 Image of sonar su data with perc 37 O ximage EJ ximage O ximage O ximage MEE 0 5 Figure 3 4 Comparison of the default hsv0 hsv2 and hsv7 colormaps Rendering these plots in grayscales emphasizes the location of the bright spot in the colorbar 38 400 200 99 and legend 1 Figure 3 5 Image of sonar su data with perc 39 We will find perc 99 to be useful You may find that you have to apply an RMS balancing to make the data look a bit more uniform sunormalize norm med lt sonar su sunormalize norm rms suximage legend 1 sunormalize norm med lt sonar su sunormalize norm rms su
204. some of these characteristics as part of processing but we have common assumptions about our data that we make or that we impose on the data as a precondition for further processing The seismic source may have a time history that makes it appear complicated For example a marine air gun signature may have a bubble pulse that follows some time after the main signal Because we may think of all resulting reflections as resulting from a convolutional process involving this source wavelet complications in the source waveform 161 may cause the reflections to appear unduely complicated Another source of waveform complication in ocean seismic surveys results from a phenomeon called ghosting In addition to the direct arrival from the source there is an additional arrival that originates from a reflection path that begins at the source bounces off of the water surface and then travels to the subsurface Similarly in addition to the direct reflection from the subsurface the receiver may record an additional signal that has traveled to the ocean surface To deal with these issues we apply deconvolutional methods To apply such methods we make some physically reasonable simplifying assumptions 11 1 1 Causality If the source is an explosion or a pulse from an air gun or even a sweep from a vibrator the resulting data have the property that they are causal which is to the say that the wavelets have a definite beginning time Causality means
205. st cases of such processing occur when the output spatial coordinates on the recording surface are such that y z and yg z2 Then the remaining problem is to trade time for depth Often the symbol z is used to represent depth with z increasing positively as we go deeper into the earth Clearly special circumstances are needed for this simple case to exist Effectively such an imaging problem is one dimensional This type of problem may result from the construction of synthetic well logs from migrated seismic data or making depth sections from migrated time sections Sonar and GPR data usually have the attribute that the same piece of equipment is used as both source and receiver Furthermore this source receiver array is likely highly directional forming a beam of energy that travels straight down into the Earth Because the scattering occurs from roughness in the structures rough surface scattering of the subsurface and owing to the strength of the sources involved we may consider the reflection to have occurred from directly below the receiver To perform time depth conversion we need to know something about the velocities of the subsurface 5 3 Time to depth with suttoz depth to time with suztot The simplest approach to depth conversion is to use a simple velocity profile expressed as a function of time v t How can we have velocity as a function of time The idea is intuitive We expect the image to show reflector
206. su You should see surange lt simple su 80 traces tracl 1 80 1 80 cdp 1 80 1 80 trid 1 Sx 0 3160 0 3160 gx 0 3160 0 3160 ns 501 dt 4000 If not for example if the sx and gx fields are not set then do the following mv simple su simple orig su sushw lt simple orig su key sx gx a 0 0 b 40 40 gt simple su If the source and geophone positions are not set then the sukdmig2d program will think that all of the input data are at the position sx 0 gx 0 The method of migration is called Kirchhoff because the technique is based on an integral equation called the Kirchhoff modeling formula which is a high frequency representation of the wavefield emanating from a reflector in the form of an integral over the reflector Some people like to think of the Kirchhoff integral as describing an exploding reflector model of reflectivity In some sense the Kirchhoff migration formula 92 is an approximate inverse of the Kirchhoff modeling formula meaning that we are in an approximate sense solving for the reflectivity in the subsurface The migration integral equation may be implemented as such or as a sum an in tegral over diffractions the diffraction stack In either case the Kirchhoff migration formula may also be called a Kirchhoff WKBJ migration where the WKBJ Wentzel Kramers Brilliouin Jeffreys denotes that ray theoretic quantities are being used Most importantly traveltimes
207. su so copy the shell scripts Migtest fd Migtest ffd Migtest split Migtest pspi and Migtest gb from cwpscratch Data4 to your Temp3 directory by 108 cd scratch yourusername mkdir Temp4 cp data cwpscratch Data4 seismic3 su Temp4 cp data cwpscratch Data4 Migtest fd Temp4 cp data cwpscratch Data4 Migtest ffd Temp4 cp data cwpscratch Data4 Migtest split Temp4 cp data cwpscratch Data4 Migtest pspi Temp4 cp data cwpscratch Data4 Migtest gb Temp4 cp data cwpscratch Data4 newvelxz bin Temp4 cp data cwpscratch Data4 newvelzx bin Temp4 where these last two files are background wavespeed profiles for You may view the back ground wavespeed profile by typing ximage lt newvelzx bin n1 1500 legend 1 ximage lt newvelxz bin n1 2142 legend 1 By now you should recognize that the model is 1500 samples in depth and 2142 sam ples in the horizontal direction The file newvelzx bin and newvelxz bin are merely transposed versions of each other The reason both files are needed is that the first 4 migrations to be tested read in the wavespeed profile in constant depth slices whereas the Gaussian Beam migration reads the data in vertical slices This is an oddity of coding nothing more It will take quite a bit of time to run all of these In a classroom environment have different students run different scripts so that all scripts are run in a reasonable amount of time In each case simply type the name of the s
208. su xcur 3 title simple interp 1 amp suxwigb lt interp su xcur 3 title interp interp 1 amp by clicking and dragging the rubberbandbox we can view these datasets in detail If we zoom in on traces 10 through 20 and time values 1 05s to 1 15s in the simple su data as in Fig 7 1a In the interp su data these correspond to the traces 20 through 40 as 95 1 054 1 104 1 154 simple 1 05 10 1 18 20 b 1 155 Wj 40 simple 1 05 20 1 18 40 Figure 7 1 a The simple su data b The same data trace interpolated the interp su data You can recognize spatial aliasing in a by noticing that the peak of the waveform on a given trace does not line up with the main lobe of the neighboring traces The data in b are the same data as in a but with twice as many traces covering the same spatial range Each peak aligns with part of the main lobe of the waveform on the neighboring trace so there is no spatial aliasing 96 in Fig 7 1 b The interp 1 allows the wiggle traces to be displayed smoothly at any scale of zooming The spatial aliasing is evident in the simple su data because peak of the waveform on a given trace does not align with the main lobe of the waveform on the neighboring traces Effectively the spatial variability of the data are undersampled in simple su because the trace spacing is to coarse In real seismic data great pains are taken to have receiver s
209. such techniques to be applied In the modern world there is an increasing demand for amplitude information for the extraction of amplitude versus angle AVA also known as amplitude versus offset AVO information Balancing away all of the amplitude variability in the data is not desirable so methods that preserve amplitudes and are data driven are preferred 184 Chapter 12 Velocity Analysis on more CDP gathers and Dip Move Out Shell scripts called Velan and Velan radon are supplied in data cwpscratch Data5 These are general purpose shell script for velocity analysis The script Velan radon is designed to aid in the generation of velocities for Radon transform based mulitiple suppression but the NMO velocities picked with this script may also be used for stacking The script Velan is may be used on data that have already been multiple suppressed These shell scripts make use of a number of programs that you have used already including suvelan suwind suximage suxgraph sunmo suradon and suxwigb as well as some programs you have not used such as unisam2 velconv and smooth2 Look through the full shell scripts before attempting to run them The beginnings of both of these scripts are similar bin sh Velocity analyses for the cmp gathers Authors Dave Hale Jack K Cohen with modifications by John Stockwell NOTE Comment lines preceding user input start with set Xx Set parameters velpanel radon gain jon
210. t1 nt 2 time sample for center pivot cx1 ntr 2 trace for center pivot plane 2 dip2 4 dip of plane 2 ms trace len2 3 ntr 4 HORIZONTAL extent of plane traces ct2 nt 2 time sample for center pivot cx2 ntr 2 trace for center pivot More As with the Unix man pages typing the space bar shows the rest of the help page Each of these programs has a relatively large number of possible argument set tings The programs suxwigb and suximage both call programs named respectively xwigb and ximage Type ximage xwigb All of the setting for xwigb and ximage apply to suxwigb and suximage That is a lot of settings Correspondingly there are plotting programs that write out PostScript graphics out put for plotting supsimage psimage supswigb pswigb supswigp pswigp The SU versions of these programs call the respective programs that do not have the su prefix AAA AHH 4 2 Finding the names of programs with suname SU is big package containing several hundred programs as well as hundreds of library functions shell scripts and associated files Occasionally we would like to see the total scope of the package we are working with For an inventory of the SU programs typing 45 suname yields Seer CWP Free Programs CWPROOT usr local cwp Mains In CWPROOT src cwp main CTRLSTRIP Strip non graphic characters DOWNFORT change Fortran p
211. tests because the interpolation process put a new trace between every existing trace You may rerun the Migtest shell script changing simple su to interp su taking care to change the input parameters to correctly reflect that the number of traces is 159 and that the spacing between them is cut in half to a value of 20 7 11 1 Interpreting the result We have several ways of interpreting the presence of artifacts in the migrated simple su data First of all we may consider the spacing between the arrivals on successive traces to be so separated that each arrival acts more like a single point input so the arcs shaped artifacts represent impulse responses Second we may view the artifacts as representing terminations in the integral that is implicit in the inverse Fourier transform that is the actual mathematical operation of f k migration So these terminations give rise to endpoint contributions where the jumps in the data act like discrete limits of integration The third interpretation is that the input arrivals are spatially aliased so that the Fourier transform process thinks that certain high spatial frequencies in the data are really low spatial frequency information and are putting this information in the wrong place It is this last interpretation that we would like to investigate further 7 11 2 Recognizing spatial aliasing of data in the space time domain If we view these two input datasets in detail suxwigb lt simple
212. that 7 kAt If we define Z exp iwAt exp i then we have the so called Z transform representation of S t N SZ X Re k 1 Remarkably we are essentially able to take the Fourier transform of a digital signal by inspection All we have to do is to multiply the k th term of our sequence digital value with Z and add up the resulting terms to form the k th order polynomial in Z For a Z transform representation of a finite number of terms this is all we need to know For an infinite series such as we might obtain through a Taylor expansion of a function or through a process of long division as in the case of the geometric series We must also specify a region of convergence which is a circle in the complex plane Poles in the Z transform representation will lie outside that circle of convergence Zeros of the polynomial may lie inside the unit circle of convergence The inverse Z transform We can use the properties of the residue theorem from complex variables to get our original series back The inverse Z transform has to have the form for each term k 1 a z SZ dZ 11 2 1 where the contour C encloses the origin We can see why this is so by considering that for Z exp i and for C being a circular contour enclosing the origin Simply substituting for Z and noting that dZ iZdo j 1 Z dZ i int eeit do 2ni which shows where the division by 277 comes from The other possibility is to consider for n gt 0 r f
213. that there can be no signal before time zero or in the case of propagating arrivals there can be no arrivals before the shortest traveltime determined by the velocity function for the medium This of course is a basic principle of physics and should not be a surprise Many processes that change the frequency spectrum also will tend to cause a distortion in the waveform resulting in signals that may appear earlier than the time predicted by the wavespeed Sometimes we deliberately change the phase characteristics of the wavelets in the data as to make the wavelet symmetric about the expected arrival time Such signals are called zero phase waveforms If the data are then processed to appear to be similar to sinc functions which is to say symmetric zero phase signals then the reflectors will occur at times of the peaks of these bandlimited delta functions 11 1 2 Minimum phase A signal which has a definite time of beginning and also has the majority of its energy in the beginning part of the waveform is called a minimum phase wavelet Again many linear or mildly nonlinear physical processes will produce signals that have this property If we think of deconvolution as division in the frequency domain then we can see that minimum phase waveforms have a special property The Z transform representation of a minimum phase signal is a polynomial that has its zeros inside the unit circle in the complex Z plane Because the Z transform repre
214. the impulse response or the transfer function 11 2 1 Convolution of a wavelet with a reflectivity series The simplest model of a seismic trace is to consider the notion of the reflectivity series The idea is simple The world is assumed to first order to consist of simple reflectors each with its own reflection coefficient Ry for the k th reflector Seismic waves represented as rays that travel from the source to each reflector and back taking T for the two way traveltime from source to the k th reflector and back to receiver There is only single scattering assumed in this ideal world so there are no multiples The simplest seismogram that could be recorded would then be a collection of spikes of each of a respective height Rpg having values which could be positive or negative arriving at the respective time Tk N S t 5 R 6 t Tk k 1 where 6 t Tp is the Dirac delta function Think of this as a spike that only turns on when t T 0 This series is called the reflectivity series a popular notion in exploration seismology If we want to make a seismogram then we would convolve a wavelet W t with the reflectivity series to form a seismogram The Z transform If we take the causal Fourier transform of our reflectivity series S t Sw i S t dt f XO Rk t Tee dt 0 0 k 1 167 N R e 5y Heer 1 k 1 I Ms k where we note that the time sampling interval is a constant At such
215. the channels away from the walls The ocean level was maintained through a variable discharge siphon located in the op posite corner of the basin Though we imposed a gradual base level rise in order to simulate subsidence the shoreline maintained a constant position through the experiment Dr Tono goes on to describe the experimental layout The outgoing pulse is generated with a Prototype JRS DPR300 Pulser Receiver which drives a 900 volt square pulse into the transducer It is set to a pulse receive frequency of 100 Hz with an input gain of 30 dB in echo mode The high pass filter is set at 20 KHz and the low pass filter at 10 MHz A Gage Applied Compuscope 1602 digitizer computer card 16 Bit 2 Channel card with acquisition memory of 1 Million samples is used to perform the A D conversion and the data is displayed on a computer screen by means of GageScope 3 50 It is digitally recorded on the computer hard disk A sample rate of 2 5 MS s is chosen Nyquist frequency 1 25 MHz It is then re formatted to SEG Y and processed with Seismic Unix The data were acquired with a 5mm shotpoint and station interval zero offset and 1cm separation between lines In the directory data cwpscratch Datal you will find a number of JPEG format files depicting the experimental setting described by Dr Tono The file dsc01324 su is an SU format file version of the image DSC01324 JPG cropped to remove parts of the image t
216. the selfdoc for the same program suplane 46 As the number of SU programs you come in contact increases you will find it useful to continually be referring to the listing from suname The sudoc feature is an alternative to Unix man pages The database of sudocs is captured from the actual selfdocs in the source code automatically via a shell script so these do not go out of step with the actual code the way a separately written man page might 4 3 Lab Activity 3 Exploring the trace header structure You may have noticed that the plotting programs seem to know a lot about the data you have been viewing Yet you have never been asked to give the number of samples per trace or the number of traces For example suximage lt sonar su perc 99 amp shows a plot without being told the dimensions of the data But how did the program know the number of traces and the number of samples per trace in the data The program knows because this and all other SU programs read information from a header that is present on each seismic trace 4 3 1 What are the trace header fields sukeyword If you type sukeyword o you will obtain a listing of the file segy h which defines the SU trace header format The term segy is derived from SEG Y a popular data exchange standard established by the Society of Exploration Geophysicists SEG in 1975 and later revised in 2005 The SU trace header is largely the same as that defined for
217. to capture the side lobes of the wavelet that appears in the center of each of the resulting traces This waveform is the autocorre lation waveform For data that are dominated by spikes or are spectrally white the autocorrelation waveform would also be a spike We pick the width of the autocorrelation waveform In the case of our example this is between approximately 0 0667 and 0 1340 seconds making the width of the autocorre lation waveform approximately 0673 seconds We apply supef setting this value as the value of maxlag supef maxlag 0673 lt fake su suxwigb perc 99 The data are made more spike like by the operation Try different values of maxlag to see what the effect of changing this parameter is Also you might want to view the change in the autocorrelation waveform supef lt fake su maxlag 0673 suacor ntout 51 suxwigb perc 90 Try the same operations on CDP 265 suacor ntout 51 lt gain jon 1 cdp 265 su suxwigb perc 99 The autocorrelation waveform is a sinc like function We define the width of this waveform to be the window of time just large enough to include the side lobes on each side of the main lobe If we measure the time from the beginning to the end of the autocorrelation waveform which is to say about 169 seconds to about 247 seconds see that the width is about 078 seconds Your values may differ This is the value of maxlag that we will set in supef supef lt gain jon 1 cdp 265 s
218. try can be one of the biggest headaches in preparing data for processing Vendors may remap the SEG Y header fields both in terms of the meaning of the header field and with respect to the data type Obtaining a header map of data when you obtain seismic data can prevent confusion and save you a lot of work When receiving data on tape remember that tape reading is more of an art than a science It is best to ask for data in a format you can use rather than allow someone else to dictate that format to you 58 Chapter 5 Lab Activity 4 Migration Imaging as depth conversion Geophysical imaging often called migration in the seismic context is an example of a general topic called inverse scattering imaging Simply stated inverse scattering imaging is the process of making pictures with echos We have all encountered examples of this in our daily lives Our eyes operate by making images of the world around us from scattered light Medical ultrasound uses the echos of high frequency sound waves to image structures within the human body Ultrasound is also used in an industrial setting for nondestructive testing NDT Seismic prospectors look for oil using the echos of seismic waves Earthquake seismologists determine the internal structure of the deep earth with echos of waves from earthquakes Near surface investigators use the echos of ground penetrating radar waves to image objects in the shallow subsurface
219. typ sweep type code 1 linear 2 cos squared 3 other short stas sweep trace length at start in ms short stae sweep trace length at end in ms short tatyp taper type 1l linear 2 cos 2 3 other short afilf alias filter frequency if used short afils alias filter slope short nofilf notch filter frequency if used 53 short short short short short short short nofils notch filter slope lcf low cut frequency if used hcf high cut frequency if used lcs low cut slope hcs high cut slope year year data recorded day day of year short hour hour of day 24 hour clock short minute minute of hour short sec second of minute short timbas time basis code 1 local 2 GMT 3 other short trwf trace weighting factor defined as 1 2 N volts for the least significant bit short grnors geophone group number of roll switch position one short grnofr geophone group number of trace one within original field record short grnlof geophone group number of last trace within original field record short short gaps gap size total number of groups dropped otrav overtravel taper code 1 down or behind 2 up or ahead 54 cwp local assignments float d1 sample spacing for non seismic data float f1
220. u maxlag 078 suxwigb xcur 3 To see how well this has spiked the data we may view the autocorrelation waveform with suacor supef lt gain jon 1 cdp 265 su maxlag 078 suacor ntout 51 suxwigb perc 90 which should show that the autocorrelation waveform is now a spike Again we may vary the value of maxlag to see the effect of changing this parameter 174 11 3 2 Multiple suppression by Wiener filtering gapped prediction error filtering We now seek to eliminate multiples by prediction error filtering also known as predictive deconvolution Predictive decon relies on the minimum phase assumption and the notion that the data are repetitious By selecting the appropriate combination of minlag and maxlag defining the Wiener filter we can eliminate repetitions in the data such as those caused by multiples This is known as gapped predictive decon in the paralance of the geophysical community We begin by spiking our fake water pegleg su which are our data with water bottom and pegleg multiples supef lt faketwater pegleg su maxlag 0673 suxwigb perc 99 xcur 2 We then view the autocorrelation of the data in a broader window choosing ntout 1024 samples in the output The idea is to look for repetitions in the autocorrelation supef lt fake twater pegleg su maxlag 0673 suacor ntout 1024 suxwigb perc 90 What we are looking for are repetitions in the autocorrelation We know that the two way traveltime for the w
221. u will see the input data for a wavespeed profile building program called unif2 The contents of this file define two boundaries in a velocity model The data for the two boundaries is separated by the values ae 99999 The values in the column on the left are horizontal positions and the values on the right are depths This model defines the same simple syncline model seen in Fig 6 1 We now look at the contents of the shell script XSyncline more XSyncline bin sh Shell script to build velocity profiles with unif2 input parameters 71 modelfile syncline unif2 velfile syncline bin n1 200 n2 400 di 10 d2 10 use unif2 to build the velocity profile unif2 lt modelfile method i ninf 2 nx n2 nz n1 vO00 1000 2000 ninf 1 method spline gt velfile view the velocity profile on the screen ximage lt velfile wbox 400 hbox 200 ni n1 n2 n2 d1i d1 d2 d2 wbox 800 hbox 400 legend 1 title Syncline model labeli depth m label2 distance m units m s amp provide input for sufdmod2 xs 1000 zs 10 hsz 10 vsx 1000 verbose 2 vsfile vseis su ssfile sseis su hsfile hseis su tmax 3 0 mt 10 label1 Depth m label2 Distance m perform finite difference acoustic modeling to generate data for a single shot in the sufdmod2 lt velfile nz n1 dz d1 nx n2 dx d2 verbose 1 xs xs zs zs hsz hsz vsx vsx hsfile hsfile vsfile vsfile ssfile ssfile verbose verbose tmax tmax abs 1 1 1 1
222. uld be the perfect semblance plot In Figure 10 6 b we see the semblance plot for these same synthetic data contaminated with water bottom multiples which are the arrivals at 1500m s the speed of sound in water arriving at intervals 0 5s reflecting the two way traveltime in the water layer The pegleg multiples are added for select events and the semblance is plotted in Figure 10 6 c The pegleg multiples are reverberations spaced at 0 5s but with decreasing velocity As we can see as the pegleg multiples bounce repetitively in the water layer their velocity approaches the water speed of 1500m s We may generate corresponding Radon i e 7 p domain plots of these data via sunmo vnmo 1500 lt fake su suradon offref 3237 interoff 262 igopt 2 choose 0 pmin 2000 pmax 2000 dp 8 depthref 1000 suximage perc 99 title fake label1 tau label2 p sunmo vnmo 1500 lt faketwater su suradon offref 3237 interoff 262 igopt 2 choose 0 pmin 2000 pmax 2000 dp 8 depthref 1000 suximage perc 99 title faket water label1 tau label2 p sunmo vnmo 1500 lt faketwater pegleg su suradon offref 3237 interoff 262 igopt 2 choose 0 pmin 2000 pmax 2000 dp 8 depthref 1000 suximage perc 99 title faket twatert tpegleg labeli tau label2 p Finally in Figure 10 7 we see the corresponding panels in the Radon r p or slant stack domain The data have been NMO corrected to flatten arrivals traveling at the water speed Here
223. uld type atq 2 Wed Oct 13 01 00 00 2009 a yourusername The first number is a number designating a job number the rest of the fields show the date and time of execution and your username on the system You are now free to log out and go on with other tasks The system will send you an email message However if you have special messages that you want the script to email you you might have lines like these in your script echo Run completed usr bin mail s Status of job myusername mymail mines edu Here the echo command is sending a message to your CSM email address with Subject line saying Status of job and message contents saying Run completed You could have other information emailed to you If your script fails the system will email you with that information as part of the standard operation of at Before using this for anything big try rerunning the Migtest scripts that you ran for Homework Problem 4 as at jobs For example try running Migtest gb at 1 am tomorrow morning cd Temp4 cp Migtest gb cp newvelzx bin cp seismic3 su at tam tomorrow Migtest gb A A A A SH 159 Then log out and check your email in the morning Make sure that it works as desired before running bigger jobs The script should run as before and you should get an email with the same screen output that you saw when you ran this on the commandline in the lab There will be an error because an at job cannot open an X wind
224. uradon lt junk su offref 3237 interoff 262 pmin 2000 pmax 2000 dp 8 choose 1 igopt 2 pmula 800 pmulb 47 depthref 1000 sunmo vnmo 1500 invert 1 gt junki su 147 p n 500 1000 500 0 500 CMP 265 in the tau p domain Figure 10 8 CMP 265 NMO corrected with vnmo 1500 displayed in the Radon trans form 7 p domain Compare this figure with Figure 10 2 The repetition indicates multiples 148 The values of pmula 800 and pmulb 47 define the beginning and ending p values of the location of the multiples The program smoothly suppresses items to the right of the line defined by tau 6 0 and p 800 and tau 0 and p 47 The invert 1 causes sunmo to apply inverse NMO which is an approximate inversion of the NMO correction to the waterspeed back to the original data Some muting occurs in the right place a lucky accident as a result of the inverse NMO To see what we obtain from the application of this filter suxwigb lt junkl su title data after multiple suppression amp To see what we removed from the data we run suradon in the choose 2 mode suradon lt junk su offref 3237 interoff 262 pmin 2000 pmax 2000 dp 8 choose 2 igopt 2 pmula 800 pmulb 47 depthref 1000 sunmo vnmo 1500 invert 1 gt junki su which gives us an estimate of what was suppressed in the data suxwigb lt junk2 su title multiples that were suppressed amp A small shell script called Radon test located in data cwpscr
225. use the Radon transform is invertible one or more of the dipping lines now represented as points can be surgically removed and the inverse transform performed to yield filtered data with one or more of the planes removed This is different from dip filtering in that there is no frequency domain band limiting effect Suppose we had applied NMO to the data so that the most steeply dipping items were the items that we wanted to remove These steeply dipping arrivals correspond to the arrivals near p 1000 in the Radon transformed plot We could perform a Radon transform surgically remove this arrival and then perform an inverse Radon transform to reconstruct the data Fortunately the suradon program also works as a filter in the Radon domain If we want to get rid of everything for p gt 600 then we need merely run the program again with the choose option Instead of a Radon transformed dataset the output this time is the data panel in the time domain with desired items removed Setting the values of pmula 600 and pmulb 600 defines a vertical line in the Radon domain at p 600 The suradon program applies a smooth filter to remove contributions to the right of a vertical line to the right of this line suradon lt suplanedata su igopt 3 interp 4 choose 1 depthref 1000 pmula 600 pmulb 600 interoff 0 offref 1190 pmin 1000 pmax 1000 gt filtered su suximage lt filtered su labell tau s label2 slope index title suplane data filtered in
226. utcome of our deconvolution So if we are deconvolving our data with a known wavelet then many processes assume that we have a non zero spectrum to avoid division by zero Many algorithms get around this by including an additive noise parameter which is usually a small number added to prevent division by zero This parameter is called pnoise in SU programs that do deconvolution The term white spectrum refers to the allusion of white light being composed of a full visual spectrum of frequencies 163 11 1 4 Lab Activity 18 Frequency filtering The simplest yet one of the most important spectral methods is simple frequency filtering If we look at the spectra of the traces in CMP gather 265 suspecfx lt gain jon 1 cdp 265 su suxgraph title spectra title spectra label1 frequency label2 amplitude amp we see that the data have most of their frequency spectral values between 5 Hz and 80 Hz The rest can be considered to be noise which could be boosted by futher processing steps Applying simple frequency filtering with sufilter we see in the frequency domain sufilter lt gain jon 1 cdp 265 su f 0 5 70 80 amps 0 1 1 0 suspecfx suxgraph title spectra title spectra label1 frequency label2 amplitude amp that the data are truncated It is up to the user to figure out the full range of frequencies in the data that are to be kept It may take some experimentation with further processing steps to
227. veloped Two of these general transformations are called transformation to zero offset TZO or migration to zero offset MZO As the names suggest data are tranformed through a migration like operation to synthetic zero offset data The motivation for developing such operations follow from the computation cost of doing full prestack migrations Alternately these techniques can be applied as velocity analysis techniques Indeed NMO or NMO followed by DMO are really first approaches to transformation to zero offset 12 5 1 Implementing DMO Note you need storage space for one or more copies of the full dataset if you want to try this For example we may NMO correct the data via sunmo par nmovel par lt radon gain jon 1 cdp su gt nmo radon gain jon 1 cdp su susort offset sx lt nmo radon gain jon 1 cdp su gt nmo radon gain jon 1 co su where it is important to note that it is not a good idea to try to use pipes with susort The co in the extension name indicates that these data are now in common offset gathers We may then perform dip moveout processing with sudmofk This program requires an average set of tdmo vdmo pairs which may are an average set of times and velocities taken from nmovel par and which are copied into a file named say dmofk par sudmofk par dmofk par cdpmin 1 cdpmax 2142 dxcdp 12 5 noffmix 7 lt nmo radon gain jon 1 co su gt dmofk nmo radon gain jon 1 co su Here we noffmix 7 is chosen fo
228. verything else revealing only the reflector Does this method work Yes but it is subject to interference errors if the data are not densely sampled in space Because a point at 7 represents an impulse in the x t space corresponding 75 y Figure 6 5 The light cone representation of the constant velocity solution of the 2D wave equation Every wavefront for both positive and negative time t is found by passing a plane parallel to the x z plane through the cone at the desired time t We may want to run time backwards for migration Figure 6 6 The light cone representation for negative times is now embedded in the x z t cube A seismic arrival to be migrated at the coordinates T is placed at the apex of the cone The circle that we draw on the seismogram for that point is the set of points obtained by the intersection of the cone with the t 0 plane 76 Distance km 1 2 3 4 5 6 yee J aml g ee alli Diffraction Figure 6 7 Hagedoorn s method of graphical migration applied to the diffraction from a point scatterer Only a few of the Hagedoorn circles are drawn here but the reader should be aware that any Hagedoorn circle through a diffraction event will intersect the apex of the diffraction hyperbola circle drawn in Hagadoorn s method m
229. vision process 11 2 5 Cross and auto correlation A related mathematical operation to convolution is the cross correlation The cross correlation of two functions is the multiplication of one function by the com plex conjugate of the other in the frequency domain Here we represent the cross correlation by the symbol xcor which for digital data is the serial multiplication of the discrete representations of A t and B t We write this as in the Fourier domain as A t xcor B t Je a w b w e dw Or I 260 1 sex me eis b tt f oe fa w e wW We can see that the auto correlation is the product of a function with its own complex conjugate in the frequency domain A t xcor A t T a w a w e dw Thus the frequency domain representation of the autocorrelation of our waveform is given by the w w which appears the denominator of the frequency domain form of the deconvolution and in the Fourier transform representation of inverse wavelet W t Whether deconvolution is performed in the time domain or in the frequency domain the common elements of the auto correlation the noise or whitening parameter and the wavelet W t are present 171 Z transform view of cross correlation Given the Z transform representations of two signals B Z and A Z N N A Z Sl aZ and B Z D2 Z k 1 1 we represent the complex conjugate B Z as the same series but with terms represented by the negative powers of Z N BZ
230. vity 14 CMP CDP Gathers 9 0 3 SOM PANG PAIN oe 3 2 e x od Ree de a hee Rais 2 Oe 9 6 4 Viewing the headers o aise an Bde oe Bde ei te gies Oo gt Stacking Chart saii Ae Pla hehe Sled oS he oe a 9 6 6 Capturing a Single CMP gather a be a be ae ee 9 7 Quality control through raw CV and brute stacks 9 7 1 Lab Activity 15 Raw Stacks CV Stacks and Brute Stacks 9 8 Homework 5 Due Thursday 27 Sept 2012 prior to 9 00AM 9 8 1 Are we done with gaining 2 5 g hid die ee eae Se eae A 104 9 9 Concluding Remarks lt 2252 250 A ood a See Ek gee da Se eh 133 10 Velocity Analysis Preview of Semblance and noise suppression 135 10 0 1 Creative use of NMO and Inverse NMO 2 138 10 1 The Radon or 7 p Transform 4 g 4eees wie ewe ho ee et 138 10 1 1 How filtering in the Radon domain differs from f k filtering 142 10 1 2 Semblance and Radon fora CDP gather 142 10 2 Multiple suppression Lab Activity 17 Radon transform 147 11 10 2 1 Homework assignment 6 Due Thursday 4 Oct 2012 before 9 00am 150 10 2 2 Are we finished with multiple suppression and velocity analysis 152 10 3 Mutine T visit d a ns pa as e Sette Oh bt ete a t Cle fm A at i 152 10 3 1 Thestr teh mute rse aaee ar BRE ae RE ik Se Ow g 152 10 3 2 Muting specific arrivals o oaa Bin as 4 By he he wR 154 10 3 3 Lab Activity 16 muting the data ccc fle hs ae
231. ximage legend 1 perc 99 Again these commands are written as one long line and are broken here to fit on the page You may zoom in on regions of the plot you find interesting If you put both the median normalized and simple perc 99 files on the screen side by side there are differences but these may not be striking differences The program suximage has a feature that the user may change colormaps by pressing the h key or the r key Try this and you will see that the selection of the colormap can make a considerable difference in the appearance of the image Even with the same data the colormap For example in Figure 3 7 we see the result of applying median normalization We might consider applying sunormalize directly to the seismic data suximage lt seismic su wbox 250 hbox 600 cmap hsv4 clip 3 title no median amp compared with applying the median normalization sunormalize norm med lt seismic su suximage wbox 250 hbox 600 cmap hsv4 clip 3 title median filtering amp This result looks bizarre because the traces individually have different median values and consequently have different ranges of amplitudes An improved picture may be obtained by applying an RMS balancing to the traces after they have been median filtered via sunormalize norm med lt seismic su sunormalize norm rms suximage wbox 250 hbox 600 cmap hsv4 clip 3 title median filtering amp In each of these examples the line is brok
232. y that which a seismic processor of the late 1970s would have experienced on a vastly slower more expensive and more difficult to use processing platform 11 1 2 Unix and Unix like operating systems The Unix operating system as well as any other Unix like operating system which includes the various forms of Linux UBUNTU Free BSD Unix and Mac OS X is commonly used in the exploration seismic community Consequently learning aspects this operating system is time well spent Many users may have grown up with a point and click environment where a given program is run via a graphical user interface GUI featuring menus and assorted windows Certainly there are such software applications in the world of commercial seismic processing but none of these are inexpensive and none give the user access to the source code of the application There is also an expert user level of work where such GUI driven tools do not exist however and programs are run from the commandline of a terminal window or are executed as part of a processing sequence in shell script In this course we will use the open source CWP SU Seismic Unix called simply Seis mic Unix or SU seismic processing and research environment This software collection was developed largely at the Colorado School of Mines CSM at the Center for Wave Phenomena CWP with contributions from users all around the world The SU soft ware package is designed to run under any Unix or U
233. ystem from the users has been a priority from day one In none of these examples have we used a browser yet there are browsers available on most Unix systems There is no fundamental problem with using a browser with the exception that you have to take your hands off the keyboard to use the mouse The browser will not tell you where you are located within a terminal window If you must use a browser use column view rather than icon view as we will have many levels of nested directories to navigate 1 9 Scratch and Data directories Directories with names such scratch and data are often provided with user write permission so that users may keep temporary files and data files out of their home direc tories Like scratch paper a scratch directory is usually for temporary file storage and is not backed up Indeed on any computer system there may be scratch directories that are not called by that name Also such directories may be physically located on the specific machine were you are seated and may not be visible on other machines Because the redundancy of backups require extra storage most system administrators restrict the amount of backed up space to a relatively small area of a computer system To restrict user access quotas may be imposed that will prevent users from using so much space that a single user could fill up a disk However in scratch areas there usually are no such restriction so it is prefe

Download Pdf Manuals

image

Related Search

Related Contents

Central Vacuum Cleaners Direct - User Guide  Samsung 713BMPLUS Uporabniški priročnik  

Copyright © All rights reserved.
Failed to retrieve file