⭐ 欢迎来到虫虫下载站! | 📦 资源下载 📁 资源专辑 ℹ️ 关于我们
⭐ 虫虫下载站

📄 paper.tex

📁 国外免费地震资料处理软件包
💻 TEX
📖 第 1 页 / 共 3 页
字号:
\end{equation}It is easy to show algebraically that estimate~(\ref{eqn:dinv1}) is equivalent to estimate~(\ref{eqn:minv1}) under the condition\begin{equation}\mathbf{C} = \mathbf{P}\,\mathbf{P}^T = \left(\mathbf{D}^T\,\mathbf{D}\right)^{-1}\;,\label{eqn:cond}\end{equation}where we need to assume the self-adjoint operator$\mathbf{D}^T\,\mathbf{D}$ to be invertible.  To prove the equivalence, consider the operator\begin{equation}\mathbf{G} = \mathbf{L}^T\,\mathbf{L}\,\mathbf{C}\,\mathbf{L}^T + \epsilon^2\,\mathbf{L}^T\;,\label{eq:g}\end{equation}which is a mapping from the data space to the model space.Grouping the multiplicative factors in two different ways, we canobtain the equality\begin{equation}\mathbf{G} = \mathbf{L}^T\,\left(\mathbf{L}\,\mathbf{C}\,\mathbf{L}^T + \epsilon^2\,\mathbf{I}\right) = \left(\mathbf{L}^T\,\mathbf{L} + \epsilon^2 \mathbf{C}^{-1}\right)\,\mathbf{C}\,\mathbf{L}^T\;,\label{eq:gmd}\end{equation}or, in another form,\begin{equation}\mathbf{C}\,\mathbf{L}^T\,\left(\mathbf{L}\,\mathbf{C}\,\mathbf{L}^T + \epsilon^2\,\mathbf{I}\right)^{-1} \equiv \left(\mathbf{L}^T\,\mathbf{L} + \epsilon^2\,\mathbf{C}^{-1}\right)^{-1}\,\mathbf{L}^T\;.\label{eq:md}\end{equation}The left-hand side of equality (\ref{eq:md}) is exactly the projectionoperator from formula (\ref{eqn:dinv1}), and the right-hand side is theoperator from formula (\ref{eqn:minv1}).This proves the legitimacyof the alternative data-space approach to data regularization: themodel estimation is reduced to a least-squares minimization of thespecially constructed compound model $\hat{\mathbf{p}}$ under theconstraint~(\ref{eqn:r}).%\begin{latexonly}We summarize the differences between the model-space and data-spaceregularization in Table 1.\tabl{table}{Comparison between model-space and data-space regularization}{\centering\begin{tabular}{|l||p{2in}|p{2in}|} \hline  \multicolumn{3}{c}{\rule[-3mm]{0mm}{8mm}} \\  \hline \rule[-3mm]{0mm}{8mm}   & \emph{Model-space} & \emph{Data-space} \\ \hline \hline  \rule[-8mm]{0mm}{17mm} effective model & $\mathbf{m}$ & $\hat{\mathbf{p}} =  \left[\begin{array}{c} \mathbf{p} \\ \mathbf{r} \end{array}\right]$ \\ \hline  \rule[-8mm]{0mm}{17mm} effective data & $\hat{\mathbf{d}} =  \left[\begin{array}{c} \mathbf{d} \\ \mathbf{0}    \end{array}\right]$ & $\mathbf{d}$ \\ \hline \rule[-8mm]{0mm}{17mm}  effective operator & $\mathbf{G_m} = \left[\begin{array}{c} \mathbf{L} \\      \epsilon \mathbf{D} \end{array}\right]$ & $\mathbf{G_d} =  \left[\begin{array}{cc} \mathbf{LP} & \epsilon \mathbf{I} \end{array}\right]$ \\  \hline \rule[-3mm]{0mm}{8mm} optimization problem & minimize  $\hat{\mathbf{r}}^T \hat{\mathbf{r}}$, \newline where \newline  \rule[-3mm]{0mm}{8mm} $\hat{\mathbf{r}} = \hat{\mathbf{d}} - \mathbf{G_m m}$ &   minimize  $\hat{\mathbf{p}}^T \hat{\mathbf{p}}$ \newline under the constraint  \newline \rule[-3mm]{0mm}{8mm} $\mathbf{G_d} \hat{\mathbf{p}} = \mathbf{d}$ \\  \hline \rule[-5mm]{0mm}{11mm} formal estimate for $\mathbf{m}$ &  $\left(\mathbf{L}^T \mathbf{L} +     \epsilon^2 \mathbf{C}^{-1}\right) \mathbf{L}^T \mathbf{d}$, \newline  \rule[-5mm]{0mm}{11mm} where $\mathbf{C}^{-1} = \mathbf{D}^T \mathbf{D}$ &   $\mathbf{C L}^T (\mathbf{L C L}^T  + \epsilon^2 \mathbf{I})^{-1} \mathbf{d}$,   \newline \rule[-5mm]{0mm}{11mm} where $\mathbf{C}  = \mathbf{P P}^T$. \\ \hline\end{tabular}}%\end{latexonly}Although the two approaches lead to similar theoretical results, theybehave quite differently in the process of iterative optimization.  Inthe next section, we illustrate this fact with many examples and showthat in the case of incomplete optimization, the second(preconditioning) approach is generally preferable.\section{One-dimensional synthetic examples}\inputdir{oned}\plot{data}{width=6in}{ The input data (right) are  irregularly spaced samples of a sinusoid (left).}In the first example, the input data $\mathbf{d}$ were randomlysubsampled (with decreasing density) from a sinusoid (Figure\ref{fig:data}). The forward operator $\mathbf{L}$ in this case islinear interpolation. In other words, we seek a regularly sampledmodel $\mathbf{m}$ on 200 grid points that could predict the data with aforward linear interpolation. Sparse irregular distribution of theinput data makes the regularization necessary.  We applied convolutionwith the simple $(1,-1)$ difference filter as the operator $\mathbf{D}$that forces model continuity (the first-order spline). An appropriatepreconditioner $\mathbf{P}$ in this case is the inverse of $\mathbf{D}$,which amounts to recursive causal integration \cite[]{gee}. Figures\ref{fig:im1} and \ref{fig:fm1} show the results of inverseinterpolation after exhaustive 300 iterations of theconjugate-direction method. The results from the model-space anddata-space regularization look similar except for the boundaryconditions outside the data range.  As a result of using the causalintegration for preconditioning, the rightmost part of the model inthe data-space case stays at a constant level instead of decreasing tozero. If we specifically wanted a zero-value boundary condition, wecould easily implement it by adding a zero-value data point at theboundary.\plot{im1}{width=3in}{ Estimation of a continuous  function by the model-space regularization. The difference operator  $\mathbf{D}$ is the derivative operator (convolution with $(1,-1)$).}\plot{fm1}{width=3in}{Estimation of a continuous  function by the data-space regularization. The preconditioning  operator $\mathbf{P}$ is causal integration.}The model preconditioning approach provides a much faster rate of convergence.We measured the rate of convergence using the power of the model residual,defined as the least-squares distance from the current model to the finalsolution.  Figure \ref{fig:schwab1} shows that the preconditioning (dataregularization) method converged to the final solution in about 6 times feweriterations than the model regularization. Since the cost of each iteration forboth methods is roughly equal, the computational economy is evident.  Figure\ref{fig:early1} shows the final solution, and the estimates from model- anddata-space regularization after only 5 iterations of conjugate directions. Thedata-space estimate appears much closer to the final solution.\plot{schwab1}{width=3in}{Convergence of the iterative  optimization, measured in terms of the model residual.  The ``d''  points stand for data-space regularization; the ``m'' points for  model-space regularization.  }\plot{early1}{width=4in}{ The top figure is the exact  solution found in 250 iterations. The middle is with data-space  regularization after 5 iterations. The bottom is with model-space  regularization after 5 iterations.  }Changing the preconditioning operator changes the regularizationresult. Figure \ref{fig:fm6} shows the result of data-spaceregularization after a triangle smoother is applied as the modelpreconditioner. Triangle smoother is a filter with the $Z$-transform$\frac{\left(1-Z^N\right)\left(1-Z^{-N}\right)}{(1-Z)\left(1-Z^{-1}\right)}$\cite[]{Claerbout.blackwell.92}. We chose the filter length $N=6$.\sideplot{fm6}{width=3in}{Estimation of a smooth function  by the data-space regularization. The preconditioning operator  $\mathbf{P}$ is a triangle smoother.}If, instead of looking for a smooth interpolation, we want to limit the numberof frequency components, then the optimal choice for the model-spaceregularization operator $\mathbf{D}$ is a prediction-error filter (PEF). Toobtain a mono-frequency output, we can use a three-point PEF, which has the$Z$-transform representation $D (Z) = 1 + a_1 Z + a_2 Z^2$. In this case, thecorresponding preconditioner $\mathbf{P}$ can be the three-point\emph{recursive} filter $P (Z) = 1 / (1 + a_1 Z + a_2 Z^2)$. To test thisidea, we estimated the PEF $D (Z)$ from the output of inverse linearinterpolation (Figure \ref{fig:fm1}), and ran the data-space regularizedestimation again, substituting the recursive filter $P (Z) = 1/ D(Z)$ in placeof the causal integration.  We repeated this two-step procedure three times toget a better estimate for the PEF. The result, shown in Figure \ref{fig:pm1},exhibits the desired mono-frequency output. One can accommodate more frequencycomponents in the model using a longer filter.\sideplot{pm1}{width=3in}{Estimation of a mono-frequency  function by the data-space regularization. The preconditioning  operator $\mathbf{P}$ is a recursive filter (the inverse of PEF).}\subsection{Regularization after binning: missing data interpolation}\inputdir{mis1}One of the factors affecting the convergence of iterative dataregularization is clustering of data points in the output bins. Whenleast-squares optimization assigns equal weight to each data point, itmay apply inadequate effort to fit a cluster of data points withsimilar values in a particular output bin. To avoid this problem, wecan replace the regularized optimization with a less accurate but moreefficient two-step approach: data binning followed by missing datainterpolation.Missing data interpolation is a particular case of dataregularization, where the input data are already given on a regulargrid, and we need to reconstruct only the missing values in emptybins.  The basic principle of missing data interpolation is formulatedas follows \cite[]{Claerbout.blackwell.92}:\begin{quote}  A method for restoring missing data is to ensure that the restored  data, after specified filtering, has minimum energy.\end{quote}Mathematically, this principle can be expressed by the simpleequation\begin{equation}\mathbf{D m \approx 0}\;,\label{eqn:basic}\end{equation}where $\mathbf{m}$ is the model vector and $\mathbf{D}$ is the specifiedfilter.  Equation~(\ref{eqn:basic}) is completely equivalent toequation~(\ref{eqn:mreg}).  The approximate equality sign means thatequation (\ref{eqn:basic}) is solved by minimizing the squared norm(the power) of its left side.  Additionally, the known data valuesmust be preserved in the optimization scheme. Introducing the maskoperator $\mathbf{K}$, which is a diagonal matrix withzeros at the missing data locations and ones elsewhere, we can rewriteequation (\ref{eqn:basic}) in the extended form\begin{equation}\mathbf{D (I-K) m} \approx - \mathbf{D K m} = - \mathbf{D m}_k\;,\label{eqn:rigor}\end{equation}where $\mathbf{I}$ is the identity operator and $\mathbf{m}_k$represents the known portion of the data. It is important to note thatequation (\ref{eqn:rigor}) corresponds to theregularized linear system\begin{equation}\left\{\begin{array}{l}\mathbf{K m} = \mathbf{m}_k\;, \\\epsilon \mathbf{D m} \approx \mathbf{0}\end{array}\right.\label{eqn:reg}\end{equation}in the limit of the scaling coefficient $\epsilon$ approaching zero.System~(\ref{eqn:reg}) is equivalent tosystem~(\ref{eqn:main}-\ref{eqn:mreg}) with the maskingoperator~$\mathbf{K}$ playing the role of the forward interpolationoperator~$\mathbf{L}$. Setting $\epsilon$ to zero implies associatingmore weight on the first equation in (\ref{eqn:reg}) and using thesecond equation only to constrain the null space of the solution.Applying the general theory of data-space regularization fromthe previous section, we can immediately transform system(\ref{eqn:reg}) to the equation\begin{equation}\mathbf{K P p} \approx  \mathbf{m}_k\;,\label{eqn:prec2}\end{equation}where $\mathbf{P}$ is a preconditioning operator and $\mathbf{p}$ is thepreconditioning variable, connected with $\mathbf{m}$ by the simplerelationship~(\ref{eqn:prec}).  According to equations~(\ref{eqn:dinv1})and~(\ref{eqn:minv1}) from the previous section, equations(\ref{eqn:prec2}) and (\ref{eqn:rigor}) have exactly the samesolutions if condition~(\ref{eqn:cond}) is satisfied.  If $\mathbf{D}$ isrepresented by a discrete convolution, the natural choice for$\mathbf{P}$ is the corresponding deconvolution (inverse recursivefiltering) operator:\begin{equation}\mathbf{P}  = \mathbf{D}^{-1}\;.\label{eqn:decon}\end{equation}We illustrate the missing data problem with a simple 1-D synthetic datatest taken from \cite{gee}.  Figure~\ref{fig:mall} shows theinterpolation results of the unpreconditioned technique with twodifferent filters $\mathbf{D}$. For comparison with the preconditioned scheme, wechanged the boundary convolution conditions from internal to truncatedtransient convolution. As in the previous example, the system wassolved with a conjugate-gradient iterative optimization.\plot{mall}{width=6in,height=4in}{Unpreconditioned  interpolation with two different regularization filters. Left plot:  the top shows the input data; the middle, the result of  interpolation; the bottom, the filter impulse response. The right  plot shows the convergence process for the first four iterations.}As depicted on the right side of the figures, the interpolationprocess starts with a ``complicated'' model and slowly extendsit until the final result is achieved. Preconditioned interpolation behaves differently (Figure\ref{fig:sall}). At the early iterations, the model is simple. As theiteration proceeds, new details are added into the model.  After a

⌨️ 快捷键说明

复制代码 Ctrl + C
搜索代码 Ctrl + F
全屏模式 F11
切换主题 Ctrl + Shift + D
显示快捷键 ?
增大字号 Ctrl + =
减小字号 Ctrl + -