Natural signals are continues, therefore, digitizing is an essential task enabling us to use computing tools to process them. According to the Nyquist/Shannon sampling theory, the sampling frequency must be at least twice the maximum frequency contained in the signal which is being sampled; otherwise, some high frequencies may be aliased and result in a bad reconstruction. The Nyquist sampling rate makes it possible to reconstruct the original signal exactly from its acquired samples. To enhance the efficiency of sampling process, a procedure is to use a high sampling rate. But the huge volume of generated data by this approach is a major challenge in many fields, like seismic exploration, and moreover, sometimes the sampling equipment cannot handle the broad frequency band. Seismic data acquisition includes sampling in time and spatial directions of a waveform that is generated by some sources like dynamite. Sampling should be done according to a regular pattern of receivers. Nevertheless, generally due to some acquisition obstacles seismic data sets are irregularly sampled in spatial direction(s). This irregularity causes a low quality seismic images that contain artifacts and missing traces. One of the approaches that have been developed to deal with this defect is interpolation of the acquired data according to a regular grid. Through the interpolation we can achieve an estimation of the fully sampled desired signal. This approach can also be as a tool to design an acquisition geometry which is sparser and results in more cost effective survey. Compressive sensing (CS) theory has been developed helping us to sample data below Nyquist sampling rate while being able to reconstruct them by considering the solution of an optimization problem. This theory claims that the signals/images that can be presented sparsely under a pre-specified basis or frame can be reconstructed accurately from a few numbers of its samples. The principle of the CS is based on the Tikhonov regularization like equation (eq. 1) which utilizes sparsifying regularization terms. In equation (1), the CS sampling operator, , contains three elements: (i) a sparsifying transform C which provides a sparse presentation of signals/images according to the used basis, (ii) measurement matrix M which for seismic issue is identity matrix, and (iii) under sampling operator S which is incoherent with sparsifying operator C. Curvelet transform contains a frame set whose elements have a great correlation with curve-like reflection events presented in seismic data and can provide a sparse presentation of seismic images. The under sampling scheme used in this paper is Jitter that allows controlling the maximum gap size between known traces. Another commonly used under sampling scheme is Gaussian random or binary random. Since under sampling appearance in frequency domain is a Gaussian random noise, the interpolation problem can be treated as a nonlinear de-noising problem. Curvelet frames are an optimal choice for this purpose. The sparsity regularization plays a leading role in CS theory. This approach has also been effectively applied on other problems like de-noising and de-convolution. There are a wide range of functions that can impose sparsity in regularization equation. The performance of these functions to interpolate an incomplete data is related to their ability in coherency with initial model properties. There are a variety of potential functions and the l1-norm is the well-known and commonly used of them. But still a comprehensive study to find out which of them is more efficient for seismic image reconstruction is necessary. This defect is because of absence of a general potential function. Here we use a general potential function which enables us to compare the efficiency of a wide range of potential functions and find the optimum one for our problem. This regularization function incudes lp-norm functions and others as its especial cases which are presented in Table 1. This general function covers both convex and non-convex regularization functions. In this paper we use the potential function to compare the efficiency of different approaches in CS algorithm. Through solving regularization problems a controversial part is setting the best regularization parameter, . Here due to redundancy of curvelet transform, assigning a proper parameter will face some difficulties. Many approaches like L-curve, Stain’s unbiased risk estimate (SURE), and generalized cross validation (GCV), face some difficulties in finding this parameter. Therefore, we inclined to use some nonlinear approaches, such as NGCV (Nonlinear GCV) and WSURE (Weighted SURE). The efficiency of the mentioned methods for estimating regularization parameter and choosing the best potential function is evaluated by considering a synthetic noisy seismic image. By under-sampling this image and removing more than 60% of its traces, the initial/observed model will be reconstructed. This imperfect image serves as our acquired seismic data. In solving equation (1) we use a forward-backward splitting recursion algorithm. Finally through this algorithm we could reach the optimum potential function and a method to estimate the regularization parameter.
Tavakoli, B., Gholami, A., & Siahkoohi, H. (2014). Seismic data interpolation via a series of new regularizing functions. Journal of the Earth and Space Physics, 40(2), 59-68. doi: 10.22059/jesphys.2014.50636
MLA
B. Tavakoli; A. Gholami; H.R. Siahkoohi. "Seismic data interpolation via a series of new regularizing functions", Journal of the Earth and Space Physics, 40, 2, 2014, 59-68. doi: 10.22059/jesphys.2014.50636
HARVARD
Tavakoli, B., Gholami, A., Siahkoohi, H. (2014). 'Seismic data interpolation via a series of new regularizing functions', Journal of the Earth and Space Physics, 40(2), pp. 59-68. doi: 10.22059/jesphys.2014.50636
VANCOUVER
Tavakoli, B., Gholami, A., Siahkoohi, H. Seismic data interpolation via a series of new regularizing functions. Journal of the Earth and Space Physics, 2014; 40(2): 59-68. doi: 10.22059/jesphys.2014.50636