HOME

TheInfoList



OR:

In signal processing, multidimensional discrete convolution refers to the mathematical operation between two functions ''f'' and ''g'' on an ''n''-dimensional lattice that produces a third function, also of ''n''-dimensions. Multidimensional discrete convolution is the discrete analog of the multidimensional convolution of functions on Euclidean space. It is also a special case of convolution on groups when the
group A group is a number of persons or things that are located, gathered, or classed together. Groups of people * Cultural group, a group whose members share the same cultural identity * Ethnic group, a group whose members share the same ethnic ide ...
is the group of ''n''-tuples of integers.


Definition


Problem statement and basics

Similar to the one-dimensional case, an asterisk is used to represent the convolution operation. The number of dimensions in the given operation is reflected in the number of asterisks. For example, an ''M''-dimensional convolution would be written with ''M'' asterisks. The following represents a ''M''-dimensional convolution of discrete signals: y(n_1,n_2,...,n_M)=x(n_1,n_2,...,n_M)* \overset *h(n_1,n_2,...,n_M) For discrete-valued signals, this convolution can be directly computed via the following: \sum_^ \sum_^...\sum_^ h(k_1,k_2,...,k_M)x(n_1-k_1,n_2-k_2,...,n_M-k_M) The resulting output region of support of a discrete multidimensional convolution will be determined based on the size and regions of support of the two input signals. Listed are several properties of the two-dimensional convolution operator. Note that these can also be extended for signals of N-dimensions. ''Commutative Property:'' x ** h = h ** x ''Associate Property:'' (x**h)**g = x**(h**g) ''Distributive Property:'' x**(h+g) = (x**h) + (x**g) These properties are seen in use in the figure below. Given some input x(n_1, n_2) that goes into a filter with impulse response h(n_1, n_2) and then another filter with impulse response g(n_1, n_2), the output is given by y(n_1, n_2). Assume that the output of the first filter is given by w(n_1, n_2), this means that: w = x ** h Further, that intermediate function is then convolved with the impulse response of the second filter, and thus the output can be represented by: y = w ** g = (x**h)**g Using the associative property, this can be rewritten as follows: y = x**(h**g) meaning that the equivalent impulse response for a cascaded system is given by: h_ = h**g A similar analysis can be done on a set of parallel systems illustrated below. In this case, it is clear that: y = (x**h)+(x**g) Using the distributive law, it is demonstrated that: y = x**(h+g) This means that in the case of a parallel system, the equivalent impulse response is provided by: h_ = h+g The equivalent impulse responses in both cascaded systems and parallel systems can be generalized to systems with N-number of filters.


Motivation and applications

Convolution in one dimension was a powerful discovery that allowed the input and output of a linear shift-invariant (LSI) system (see LTI system theory) to be easily compared so long as the impulse response of the filter system was known. This notion carries over to multidimensional convolution as well, as simply knowing the impulse response of a multidimensional filter too allows for a direct comparison to be made between the input and output of a system. This is profound since several of the signals that are transferred in the digital world today are of multiple dimensions including images and videos. Similar to the one-dimensional convolution, the multidimensional convolution allows the computation of the output of an LSI system for a given input signal. For example, consider an image that is sent over some wireless network subject to electro-optical noise. Possible noise sources include errors in channel transmission, the analog to digital converter, and the image sensor. Usually noise caused by the channel or sensor creates spatially-independent, high-frequency signal components that translates to arbitrary light and dark spots on the actual image. In order to rid the image data of the high-frequency spectral content, it can be multiplied by the frequency response of a low-pass filter, which based on the convolution theorem, is equivalent to convolving the signal in the time/spatial domain by the impulse response of the low-pass filter. Several impulse responses that do so are shown below. In addition to filtering out spectral content, the multidimensional convolution can implement edge detection and smoothing. This once again is wholly dependent on the values of the impulse response that is used to convolve with the input image. Typical impulse responses for edge detection are illustrated below. In addition to image processing, multidimensional convolution can be implemented to enable a variety of other applications. Since filters are widespread in digital communication systems, any system that must transmit multidimensional data is assisted by filtering techniques It is used in real-time video processing, neural network analysis, digital geophysical data analysis, and much more. One typical distortion that occurs during image and video capture or transmission applications is blur that is caused by a low-pass filtering process. The introduced blur can be modeled using Gaussian low-pass filtering.


Row-column decomposition with separable signals


Separable signals

A signal is said to be separable if it can be written as the product of multiple one-dimensional signals. Mathematically, this is expressed as the following: x(n_1,n_2,...,n_M) = x(n_1)x(n_2)...x(n_M) Some readily recognizable separable signals include the unit step function, and the dirac-delta impulse function. u(n_1,n_2,...,n_M)=u(n_1)u(n_2)...u(n_M) (unit step function) \delta(n_1,n_2,...,n_M)=\delta(n_1)\delta(n_2)...\delta(n_M) (dirac-delta impulse function) Convolution is a linear operation. It then follows that the multidimensional convolution of separable signals can be expressed as the product of many one-dimensional convolutions. For example, consider the case where ''x'' and ''h'' are both separable functions. x(n_1,n_2)**h(n_1,n_2)=\sum_^ \sum_^ h(k_1,k_2)x(n_1-k_1,n_2-k_2) By applying the properties of separability, this can then be rewritten as the following: x(n_1,n_2)**h(n_1,n_2)=\bigg(\sum_^ h(k_1)x(n_1-k_1)\bigg)\bigg(\sum_^h(k_2)x(n_2-k_2)\bigg) It is readily seen then that this reduces to the product of one-dimensional convolutions: x(n_1,n_2)**h(n_1,n_2)=\bigg (n_1)*h(n_1)\biggbigg (n_2)*h(n_2)\bigg/math> This conclusion can then be extended to the convolution of two separable ''M''-dimensional signals as follows: x(n_1,n_2,...,n_M)* \overset *h(n_1,n_2,...,n_M)=\bigg (n_1)*h(n_1)\biggbigg (n_2)*h(n_2)\bigg..\bigg (n_M)*h(n_M)\bigg/math> So, when the two signals are separable, the multidimensional convolution can be computed by computing n_M one-dimensional convolutions.


Row-column decomposition

The row-column method can be applied when one of the signals in the convolution is separable. The method exploits the properties of separability in order to achieve a method of calculating the convolution of two multidimensional signals that is more computationally efficient than direct computation of each sample (given that one of the signals are separable). The following shows the mathematical reasoning behind the row-column decomposition approach (typically h(n_1,n_2) is the separable signal): \begin y(n_1,n_2)&=\sum_^ \sum_^ h(k_1,k_2)x(n_1-k_1,n_2-k_2)\\ &=\sum_^ \sum_^ h_1(k_1)h_2(k_2)x(n_1-k_1,n_2-k_2)\\ &=\sum_^h_1(k_1)\Bigg \sum_^ h_2(k_2)x(n_1-k_1,n_2-k_2)\Bigg\end The value of \sum_^ h_2(k_2)x(n_1-k_1,n_2-k_2) can now be re-used when evaluating other y values with a shared value of n_2: \begin y(n_1+\delta,n_2)&=\sum_^h_1(k_1)\Bigg \sum_^_h_2(k_2)x(n_1-[k_1-\deltan_2-k_2)\Bigg.html" ;"title="_1-\delta.html" ;"title="\sum_^ h_2(k_2)x(n_1-[k_1-\delta">\sum_^ h_2(k_2)x(n_1-[k_1-\deltan_2-k_2)\Bigg">_1-\delta.html" ;"title="\sum_^ h_2(k_2)x(n_1-[k_1-\delta">\sum_^ h_2(k_2)x(n_1-[k_1-\deltan_2-k_2)\Bigg\ &=\sum_^h_1(k_1+\delta)\Bigg \sum_^ h_2(k_2)x(n_1-k_1,n_2-k_2)\Bigg\end Thus, the resulting convolution can be effectively calculated by first performing the convolution operation on all of the rows of x(n_1,n_2), and then on all of its columns. This approach can be further optimized by taking into account how memory is accessed within a computer processor. A processor will load in the signal data needed for the given operation. For modern processors, data will be loaded from memory into the processors cache, which has faster access times than memory. The cache itself is partitioned into lines. When a cache line is loaded from memory, multiple data operands are loaded at once. Consider the optimized case where a row of signal data can fit entirely within the processor's cache. This particular processor would be able to access the data row-wise efficiently, but not column-wise since different data operands in the same column would lie on different cache lines. In order to take advantage of the way in which memory is accessed, it is more efficient to transpose the data set and then axis it row-wise rather than attempt to access it column-wise. The algorithm then becomes: # Separate the separable two-dimensional signal h(n_1,n_2) into two one-dimensional signals h_1(n_1) and h_2(n_2) # Perform row-wise convolution on the horizontal components of the signal x(n_1,n_2) using h_1(n_1) to obtain g(n_1,n_2) # Transpose the vertical components of the signal g(n_1,n_2) resulting from Step 2. # Perform row-wise convolution on the transposed vertical components of g(n_1,n_2) to get the desired output y(n_1,n_2)


Computational speedup from row-column decomposition

Examine the case where an image of size X\times Y is being passed through a separable filter of size J\times K. The image itself is not separable. If the result is calculated using the direct convolution approach without exploiting the separability of the filter, this will require approximately XYJK multiplications and additions. If the separability of the filter is taken into account, the filtering can be performed in two steps. The first step will have XYJ multiplications and additions and the second step will have XYK, resulting in a total of XYJ+XYK or XY(J+K) multiplications and additions. A comparison of the computational complexity between direct and separable convolution is given in the following image:


Circular convolution of discrete-valued multidimensional signals

The premise behind the circular convolution approach on multidimensional signals is to develop a relation between the Convolution theorem and the
Discrete Fourier transform In mathematics, the discrete Fourier transform (DFT) converts a finite sequence of equally-spaced samples of a function into a same-length sequence of equally-spaced samples of the discrete-time Fourier transform (DTFT), which is a comp ...
(DFT) that can be used to calculate the convolution between two finite-extent, discrete-valued signals.


Convolution theorem in multiple dimensions

For one-dimensional signals, the
Convolution Theorem In mathematics, the convolution theorem states that under suitable conditions the Fourier transform of a convolution of two functions (or signals) is the pointwise product of their Fourier transforms. More generally, convolution in one domain (e.g ...
states that the Fourier transform of the convolution between two signals is equal to the product of the Fourier Transforms of those two signals. Thus, convolution in the time domain is equal to multiplication in the frequency domain. Mathematically, this principle is expressed via the following:y(n)=h(n)*x(n)\longleftrightarrow Y(\omega)=H(\omega)X(\omega)This principle is directly extendable to dealing with signals of multiple dimensions.y(n_1,n_2,...,n_M)=h(n_1,n_2,...,n_M)*\overset*x(n_1,n_2,...,n_M) \longleftrightarrow Y(\omega_1,\omega_2,...,\omega_M)=H(\omega_1,\omega_2,...,\omega_M)X(\omega_1,\omega_2,...,\omega_M) This property is readily extended to the usage with the
Discrete Fourier transform In mathematics, the discrete Fourier transform (DFT) converts a finite sequence of equally-spaced samples of a function into a same-length sequence of equally-spaced samples of the discrete-time Fourier transform (DTFT), which is a comp ...
(DFT) as follows (note that linear convolution is replaced with circular convolution where \otimes is used to denote the circular convolution operation of size N): y(n)=h(n)\otimes x(n)\longleftrightarrow Y(k)=H(k)X(k) When dealing with signals of multiple dimensions:y(n_1,n_2,...,n_M)=h(n_1,n_2,...,n_M)\otimes\overset\otimes x(n_1,n_2,...,n_M) \longleftrightarrow Y(k_1,k_2,...,k_M)=H(k_1,k_2,...,k_M)X(k_1,k_2,...,k_M)The circular convolutions here will be of size N_1, N_2,...,N_M.


Circular convolution approach

The motivation behind using the circular convolution approach is that it is based on the DFT. The premise behind circular convolution is to take the DFTs of the input signals, multiply them together, and then take the inverse DFT. Care must be taken such that a large enough DFT is used such that aliasing does not occur. The DFT is numerically computable when dealing with signals of finite-extent. One advantage this approach has is that since it requires taking the DFT and inverse DFT, it is possible to utilize efficient algorithms such as the Fast Fourier transform (FFT). Circular convolution can also be computed in the time/spatial domain and not only in the frequency domain.


Choosing DFT size to avoid aliasing

Consider the following case where two finite-extent signals ''x'' and ''h'' are taken. For both signals, there is a corresponding DFT as follows: x(n_1,n_2)\longleftrightarrow X(k_1,k_2) and h(n_1,n_2) \longleftrightarrow H(k_1,k_2) The region of support of x(n_1,n_2) is 0 \leq n_1 \leq P_1-1 and 0 \leq n_2 \leq P_2-1 and the region of support of h(n_1,n_2) is 0 \leq n_1 \leq Q_1-1 and 0 \leq n_2 \leq Q_2-1. The linear convolution of these two signals would be given as:y_(n_1,n_2)=\sum_ \sum_h(m_1,m_2)x(n_1-m_1,n_2-m_2)Given the regions of support of x(n_1,n_2) and h(n_1,n_2), the region of support of y_(n_1,n_2) will then be given as the following: 0\leq n_1 \leq P_1 + Q_1 - 10\leq n_2 \leq P_2 + Q_2-1Based on the regions of support of the two signals, a DFT of size N_1\times N_2 must be used where N_1 \geq \max(P_1,Q_1) and N_2 \geq \max(P_2,Q_2) since the same size DFT must be used on both signals. In the event where a DFT size larger than the extent of a signal is needed, the signal is zero-padded until it reaches the required length. After multiplying the DFTs and taking the inverse DFT on the result, the resulting circular convolution is then given by: y_(n_1,n_2)=\sum_\sum_\Bigg sum_^ \sum_^h(m_1,m_2)x(n_1-m_1-r_1N_1,n_2-m_2-r_2N_2)\Bigg/math> for (n_1,n_2) \in R_ R_\triangleq \ The result will be that y_(n_1,n_2) will be a spatially aliased version of the linear convolution result y_(n_1,n_2). This can be expressed as the following: y_(n_1,n_2)=\sum_ \sum_ y_(n_1-r_1N_1, n_2-r_2N_2)(n_1,n_2) \in R_ Then, in order to avoid aliasing between the spatially aliased replicas, N_1 and N_2 must be chosen to satisfy the following conditions: N_1 \geq P_1+Q_1-1 N_2 \geq P_2+Q_2-1 If these conditions are satisfied, then the results of the circular convolution will equal that of the linear convolution (taking the main period of the circular convolution as the region of support). That is: y_(n_1,n_2)=y_(n_1,n_2) for (n_1,n_2) \in R_


Summary of procedure using DFTs

The Convolution theorem and circular convolution can thus be used in the following manner to achieve a result that is equal to performing the linear convolution: # Choose N_1 and N_2 to satisfy N_1 \geq P_1+Q_1-1 and N_2 \geq P_2+Q_2-1 # Zero pad the signals h(n_1,n_2) and x(n_1,n_2) such that they are both N_1\times N_2 in size # Compute the DFTs of both h(n_1,n_2) and x(n_1,n_2) # Multiple the results of the DFTs to obtain Y(k_1,k_2)=H(k_1,k_2)X(k_1,k_2) # The result of the IDFT of Y(k_1,k_2) will then be equal to the result of performing linear convolution on the two signals


Overlap and add

Another method to perform multidimensional convolution is the overlap and add approach. This method helps reduce the computational complexity often associated with multidimensional convolutions due to the vast amounts of data inherent in modern-day digital systems. For sake of brevity, the two-dimensional case is used as an example, but the same concepts can be extended to multiple dimensions. Consider a two-dimensional convolution using a direct computation: y(n_1, n_2) = \sum_^ \sum_^ x(n_1 - k_1, n_2 - k_2)h(k_1, k_2) Assuming that the output signal y(n_1, n_2) has N nonzero coefficients, and the impulse response has M nonzero samples, this direct computation would need MN multiplies and MN - 1 adds in order to compute. Using an FFT instead, the frequency response of the filter and the Fourier transform of the input would have to be stored in memory. Massive amounts of computations and excessive use of memory storage space pose a problematic issue as more dimensions are added. This is where the overlap and add convolution method comes in.


Decomposition into smaller convolution blocks

Instead of performing convolution on the blocks of information in their entirety, the information can be broken up into smaller blocks of dimensions L_1xL_2 resulting in smaller FFTs, less computational complexity, and less storage needed. This can be expressed mathematically as follows: x(n_1, n_2) = \sum_^ \sum_^x_(n_1, n_2) where x(n_1, n_2) represents the N_1xN_2 input signal, which is a summation of P_1P_2 block segments, with P_1 = N_1/L_1 and P_2 = N_2/L_2 . To produce the output signal, a two-dimensional convolution is performed: y(n_1,n_2) = x(n_1, n_2) ** h(n_1, n_2) Substituting in for x(n_1, n_2) results in the following: y(n_1,n_2) = \sum_^ \sum_^x_(n_1, n_2)** h(n_1, n_2) This convolution adds more complexity than doing a direct convolution; however, since it is integrated with an FFT fast convolution, overlap-add performs faster and is a more memory-efficient method, making it practical for large sets of multidimensional data.


Breakdown of procedure

Let h(n_1, n_2) be of size M_1 \times M_2: # Break input x(n_1, n_2) into non-overlapping blocks of dimensions L_1 \times L_2. # Zero pad h(n_1, n_2) such that it has dimensions (L_1 + M_1 - 1) \times (L_2 + M_2 - 1). # Use DFT to get H(k_1, k_2). # For each input block: ## Zero pad x_(n_1, n_2) to be of dimensions (L_1 + M_1 - 1) \times (L_2 + M_2 - 1). ## Take discrete Fourier transform of each block to give X_(k_1, k_2). ## Multiply to get Y_(k_1, k_2) = X_(k_1, k_2)H(k_1,k_2). ## Take inverse discrete Fourier transform of Y_(k_1, k_2) to get y_(n_1, n_2). # Find y(n_1, n_2) by overlap and adding the last (M_1 - 1)\times (M_2 - 1) samples of y_(n_1, n_2) with the first (M_1 - 1) \times(M_2 - 1) samples of y_(n_1, n_2) to get the result.


Pictorial method of operation

In order to visualize the overlap-add method more clearly, the following illustrations examine the method graphically. Assume that the input x(n_1, n_2) has a square region support of length N in both vertical and horizontal directions as shown in the figure below. It is then broken up into four smaller segments in such a way that it is now composed of four smaller squares. Each block of the aggregate signal has dimensions (N/2) \times (N/2). Then, each component is convolved with the impulse response of the filter. Note that an advantage for an implementation such as this can be visualized here since each of these convolutions can be parallelized on a computer, as long as the computer has sufficient memory and resources to store and compute simultaneously. In the figure below, the first graph on the left represents the convolution corresponding to the component of the input x_ with the corresponding impulse response h(n_1,n_2). To the right of that, the input x_ is then convolved with the impulse response h(n_1,n_2). The same process is done for the other two inputs respectively, and they are accumulated together in order to form the convolution. This is depicted to the left. Assume that the filter impulse response h(n_1,n_2) has a region of support of (N/8) in both dimensions. This entails that each convolution convolves signals with dimensions (N/2) \times (N/8) in both n_1 and n_2 directions, which leads to overlap (highlighted in blue) since the length of each individual convolution is equivalent to: (N/2) +(N/8) -1 = (5/8)N-1 in both directions. The lighter blue portion correlates to the overlap between two adjacent convolutions, whereas the darker blue portion correlates to overlap between all four convolutions. All of these overlap portions are added together in addition to the convolutions in order to form the combined convolution y(n_1,n_2).


Overlap and save

The overlap and save method, just like the overlap and add method, is also used to reduce the computational complexity associated with discrete-time convolutions. This method, coupled with the FFT, allows for massive amounts of data to be filtered through a digital system while minimizing the necessary memory space used for computations on massive arrays of data.


Comparison to overlap and add

The overlap and save method is very similar to the overlap and add methods with a few notable exceptions. The overlap-add method involves a linear convolution of discrete-time signals, whereas the overlap-save method involves the principle of circular convolution. In addition, the overlap and save method only uses a one-time zero padding of the impulse response, while the overlap-add method involves a zero-padding for every convolution on each input component. Instead of using zero padding to prevent time-domain aliasing like its overlap-add counterpart, overlap-save simply discards all points of aliasing, and saves the previous data in one block to be copied into the convolution for the next block. In one dimension, the performance and storage metric differences between the two methods is minimal. However, in the multidimensional convolution case, the overlap-save method is preferred over the overlap-add method in terms of speed and storage abilities. Just as in the overlap and add case, the procedure invokes the two-dimensional case but can easily be extended to all multidimensional procedures.


Breakdown of procedure

Let h(n_1, n_2) be of size M_1 \times M_2 : # Insert (M_1 - 1) columns and (M_2 - 1) rows of zeroes at the beginning of the input signal x(n_1,n_2) in both dimensions. # Split the corresponding signal into overlapping segments of dimensions (L_1 + M_1 - 1)\times(L_2 + M_2 - 1) in which each two-dimensional block will overlap by (M_1 - 1) \times (M_2 - 1). # Zero pad h(n_1, n_2) such that it has dimensions (L_1 + M_1 - 1)\times(L_2 + M_2 - 1). # Use DFT to get H(k_1, k_2). # For each input block: ## Take discrete Fourier transform of each block to give X_(k_1, k_2). ## Multiply to get Y_(k_1, k_2) = X_(k_1, k_2)H(k_1,k_2). ## Take inverse discrete Fourier transform of Y_(k_1, k_2) to get y_(n_1, n_2). ## Get rid of the first (M_1 - 1)\times(M_2 - 1) for each output block y_(n_1, n_2). # Find y(n_1, n_2) by attaching the last (L_1\times L_2) samples for each output block y_(n_1, n_2).


The helix transform

Similar to row-column decomposition, the helix transform computes the multidimensional convolution by incorporating one-dimensional convolutional properties and operators. Instead of using the separability of signals, however, it maps the Cartesian coordinate space to a helical coordinate space allowing for a mapping from a multidimensional space to a one-dimensional space.


Multidimensional convolution with one-dimensional convolution methods

To understand the helix transform, it is useful to first understand how a multidimensional convolution can be broken down into a one-dimensional convolution. Assume that the two signals to be convolved are X_ and Y_, which results in an output Z_. This is expressed as follows: Z(i,j) = \sum_^\sum_^X(m,n)Y(i-m, j-n) Next, two matrices are created that zero pad each input in both dimensions such that each input has equivalent dimensions, i.e. \mathbf = \begin X & 0 \\ 0 & 0 \\ \end and \mathbf = \begin Y & 0 \\ 0 & 0 \\ \end where each of the input matrices are now of dimensions (M+K-1)\times(N+L-1). It is then possible to implement column-wise lexicographic ordering in order to convert the modified matrices into vectors, X'' and Y''. In order to minimize the number of unimportant samples in each vector, each vector is truncated after the last sample in the original matrices X and Y respectively. Given this, the length of vector X'' and Y'' are given by: l_ = (M+K-1)\times(N-1) + M l_ =(M+K-1)\times(L-1) + K The length of the convolution of these two vectors, Z'', can be derived and shown to be: l_ = l_ +l_ = (M+K-1)\times(N+L-1) This vector length is equivalent to the dimensions of the original matrix output Z, making converting back to a matrix a direct transformation. Thus, the vector, Z'', is converted back to matrix form, which produces the output of the two-dimensional discrete convolution.


Filtering on a helix

When working on a two-dimensional Cartesian mesh, a Fourier transform along either axes will result in the two-dimensional plane becoming a cylinder as the end of each column or row attaches to its respective top forming a cylinder. Filtering on a helix behaves in a similar fashion, except in this case, the bottom of each column attaches to the top of the next column, resulting in a helical mesh. This is illustrated below. The darkened tiles represent the filter coefficients. If this helical structure is then sliced and unwound into a one-dimensional strip, the same filter coefficients on the 2-d Cartesian plane will match up with the same input data, resulting in an equivalent filtering scheme. This ensures that a two-dimensional convolution will be able to be performed by a one-dimensional convolution operator as the 2D filter has been unwound to a 1D filter with gaps of zeroes separating the filter coefficients. Assuming that some-low pass two-dimensional filter was used, such as: Then, once the two-dimensional space was converted into a helix, the one-dimensional filter would look as follows: h(n) = -1, 0, ... , 0, -1, 4, -1, 0, ..., 0, -1, 0, ... Notice in the one-dimensional filter that there are no leading zeroes as illustrated in the one-dimensional filtering strip after being unwound. The entire one-dimensional strip could have been convolved with; however, it is less computationally expensive to simply ignore the leading zeroes. In addition, none of these backside zero values will need to be stored in memory, preserving precious memory resources.


Applications

Helix transformations to implement recursive filters via convolution are used in various areas of signal processing. Although frequency domain Fourier analysis is effective when systems are stationary, with constant coefficients and periodically-sampled data, it becomes more difficult in unstable systems. The helix transform enables three-dimensional post-stack migration processes that can process data for three-dimensional variations in velocity. In addition, it can be applied to assist with the problem of implicit three-dimensional wavefield extrapolation. Other applications include helpful algorithms in seismic data regularization, prediction error filters, and noise attenuation in geophysical digital systems.


Gaussian convolution

One application of multidimensional convolution that is used within signal and image processing is Gaussian convolution. This refers to convolving an input signal with the Gaussian distribution function. The Gaussian distribution sampled at discrete values in one dimension is given by the following (assuming \mu=0):G(n)=\frace^This is readily extended to a signal of ''M'' dimensions (assuming \sigma stays constant for all dimensions and \mu_1=\mu_2=...=\mu_M=0):G(n_1,n_2,...,n_M)=\frace^One important property to recognize is that the ''M'' dimensional signal is separable such that:G(n_1,n_2,...,n_M)=G(n_1)G(n_2)...G(n_M)Then, Gaussian convolution with discrete-valued signals can be expressed as the following: y(n)=x(n)*G(n) y(n_1,n_2,...,n_M)=x(n_1,n_2,...,n_M)*...*G(n_1,n_2,...,n_M)


Approximation by FIR filter

Gaussian convolution can be effectively approximated via implementation of a Finite impulse response (FIR) filter. The filter will be designed with truncated versions of the Gaussian. For a two-dimensional filter, the transfer function of such a filter would be defined as the following: H(z_1,z_2)=\frac \sum_^\sum_^G(n_1,n_2)^^ where s(r_1,r_2)=\sum_^\sum_^G(n_1,n_2) Choosing lower values for r_1 and r_2 will result in performing less computations, but will yield a less accurate approximation while choosing higher values will yield a more accurate approximation, but will require a greater number of computations.


Approximation by box filter

Another method for approximating Gaussian convolution is via recursive passes through a box filter. For approximating one-dimensional convolution, this filter is defined as the following: H(z)=\frac \frac Typically, recursive passes 3, 4, or 5 times are performed in order to obtain an accurate approximation. A suggested method for computing ''r'' is then given as the following: \sigma^2=\fracK((2r+1)^2-1) where ''K'' is the number of recursive passes through the filter. Then, since the Gaussian distribution is separable across different dimensions, it follows that recursive passes through one-dimensional filters (isolating each dimension separately) will thus yield an approximation of the multidimensional Gaussian convolution. That is, ''M-''dimensional Gaussian convolution could be approximated via recursive passes through the following one-dimensional filters: H(z_1)=\frac \frac H(z_2)=\frac \frac \vdots H(z_M)=\frac \frac


Applications

Gaussian convolutions are used extensively in signal and image processing. For example, image-blurring can be accomplished with Gaussian convolution where the \sigma parameter will control the strength of the blurring. Higher values would thus correspond to a more blurry end result. It is also commonly used in
Computer vision Computer vision is an interdisciplinary scientific field that deals with how computers can gain high-level understanding from digital images or videos. From the perspective of engineering, it seeks to understand and automate tasks that the human ...
applications such as Scale-invariant feature transform (SIFT) feature detection.


See also

*
Convolution In mathematics (in particular, functional analysis), convolution is a mathematical operation on two functions ( and ) that produces a third function (f*g) that expresses how the shape of one is modified by the other. The term ''convolution'' ...
* Kernel (image processing) *
Signal processing Signal processing is an electrical engineering subfield that focuses on analyzing, modifying and synthesizing ''signals'', such as sound, images, and scientific measurements. Signal processing techniques are used to optimize transmissions, d ...


References

{{reflist, 30em Multidimensional signal processing