The previous sections on approximating \(f\) by a finite element function \(u\) utilize the projection/Galerkin or least squares approaches to minimize the approximation error. We may, alternatively, use the collocation/interpolation method as described in the section Comparison with finite elements and interpolation/collocation. Here we shall compare these three approaches with what one does in the finite difference method when representing a given function on a mesh.
Approximating a given function \(f(x)\) on a mesh in a finite difference context will typically just sample \(f\) at the mesh points. If \(u_i\) is the value of the approximate \(u\) at the mesh point \(x_{i}\), we have \(u_i = f(x_{i})\). The collocation/interpolation method using finite element basis functions gives exactly the same representation, as shown the section Comparison with finite elements and interpolation/collocation,
How does a finite element Galerkin or least squares approximation differ from this straightforward interpolation of \(f\)? This is the question to be addressed next. We now limit the scope to P1 elements since this is the element type that gives formulas closest to those arising in the finite difference method.
The linear system arising from a Galerkin or least squares approximation reads in general
In the finite element approximation we choose \({\psi}_i ={\varphi}_i\). With \({\varphi}_i\) corresponding to P1 elements and a uniform mesh of element length \(h\) we have in the section Calculating the linear system calculated the matrix with entries \(({\varphi}_i,{\varphi}_j)\). Equation number \(i\) reads
The first and last equation, corresponding to \(i=0\) and \(i=N\) are slightly different, see the section The structure of the coefficient matrix.
The finite difference counterpart to (1) is just \(u_i=f_i\) as explained in the section Finite difference approximation of given functions. To easier compare this result to the finite element approach to approximating functions, we can rewrite the left-hand side of (1) as
Thinking in terms of finite differences, we can write this expression using finite difference operator notation:
which is nothing but the standard discretization of
Before interpreting the approximation procedure as solving a differential equation, we need to work out what the right-hand side is in the context of P1 elements. Since \({\varphi}_i\) is the linear function that is 1 at \(x_{i}\) and zero at all other nodes, only the interval \([x_{i-1},x_{i+1}]\) contribute to the integral on the right-hand side. This integral is naturally split into two parts according to (5.3):
However, if \(f\) is not known we cannot do much else with this expression. It is clear that many values of \(f\) around \(x_{i}\) contributes to the right-hand side, not just the single point value \(f(x_{i})\) as in the finite difference method.
To proceed with the right-hand side, we can turn to numerical integration schemes. The Trapezoidal method for \((f,{\varphi}_i)\), based on sampling the integrand \(f{\varphi}_i\) at the node points \(x_{i}=i h\) gives
Since \({\varphi}_i\) is zero at all these points, except at \(x_{i}\), the Trapezoidal rule collapses to one term:
for \(i=1,\ldots,N-1\), which is the same result as with collocation/interpolation, and of course the same result as in the finite difference method. For \(i=0\) and \(i=N\) we get contribution from only one element so
Simpson’s rule with sample points also in the middle of the elements, at \(x_{i+\frac{1}{2}}=(x_{i} + x_{i+1})/2\), can be written as
where \(\tilde h= h/2\) is the spacing between the sample points. Our integrand is \(g=f{\varphi}_i\). For all the node points, \({\varphi}_i(x_{j})=\delta_{ij}\), and therefore \(\sum_{j=1}^{N-1} f(x_{j}){\varphi}_i(x_{j})=f(x_{i})\). At the midpoints, \({\varphi}_i(x_{i\pm\frac{1}{2}})=1/2\) and \({\varphi}_i(x_{j+\frac{1}{2}})=0\) for \(j>1\) and \(j<i-1\). Consequently,
When \(1\leq i\leq N-1\) we then get
This result shows that, with Simpson’s rule, the finite element method operates with the average of \(f\) over three points, while the finite difference method just applies \(f\) at one point. We may interpret this as a “smearing” or smoothing of \(f\) by the finite element method.
We can now summarize our findings. With the approximation of \((f,{\varphi}_i)\) by the Trapezoidal rule, P1 elements give rise to equations that can be expressed as a finite difference discretization of
expressed with operator notation as
As \(h\rightarrow 0\), the extra term proportional to \(u''\) goes to zero, and the two methods are then equal.
With the Simpson’s rule, we may say that we solve
where \(\bar f_i\) means the average \(\frac{1}{3}(f_{i-1/2} + f_i + f_{i+1/2})\).
The extra term \(\frac{h^2}{6} u''\) represents a smoothing effect: with just this term, we would find \(u\) by integrating \(f\) twice and thereby smooth \(f\) considerably. In addition, the finite element representation of \(f\) involves an average, or a smoothing, of \(f\) on the right-hand side of the equation system. If \(f\) is a noisy function, direct interpolation \(u_i=f_i\) may result in a noisy \(u\) too, but with a Galerkin or least squares formulation and P1 elements, we should expect that \(u\) is smoother than \(f\) unless \(h\) is very small.
The interpretation that finite elements tend to smooth the solution is valid in applications far beyond approximation of 1D functions.
With a simple trick, using numerical integration, we can easily produce the result \(u_i=f_i\) with the Galerkin or least square formulation with P1 elements. This is useful in many occasions when we deal with more difficult differential equations and want the finite element method to have properties like the finite difference method (solving standard linear wave equations is one primary example).
We have already seen that applying the Trapezoidal rule to the right-hand side \((f,{\varphi}_i)\) simply gives \(f\) sampled at \(x_{i}\). Using the Trapezoidal rule on the matrix entries \(A_{i,j}=({\varphi}_i,{\varphi}_j)\) involves a sum
but \({\varphi}_i(x_{k})=\delta_{ik}\) and \({\varphi}_j(x_{k})=\delta_{jk}\). The product \({\varphi}_i{\varphi}_j\) is then different from zero only when sampled at \(x_{i}\) and \(i=j\). The Trapezoidal approximation to the integral is then
and zero if \(i\neq j\). This means that we have obtained a diagonal matrix! The first and last diagonal elements, \(({\varphi}_0,{\varphi}_0)\) and \(({\varphi}_N,{\varphi}_N)\) get contribution only from the first and last element, respectively, resulting in the approximate integral value \(h/2\). The corresponding right-hand side also has a factor \(1/2\) for \(i=0\) and \(i=N\). Therefore, the least squares or Galerkin approach with P1 elements and Trapezoidal integration results in
Simpsons’s rule can be used to achieve a similar result for P2 elements, i.e, a diagonal coefficient matrix, but with the previously derived average of \(f\) on the right-hand side.
Identical results to those above will arise if we perform elementwise computations. The idea is to use the Trapezoidal rule on the reference element for computing the element matrix and vector. When assembled, the same equations \(c_i=f(x_{i})\) arise. Exercise 19: Use the Trapezoidal rule and P1 elements encourages you to carry out the details.
The matrix with entries \(({\varphi}_i,{\varphi}_j)\) typically arises from terms proportional to \(u\) in a differential equation where \(u\) is the unknown function. This matrix is often called the mass matrix, because in the early days of the finite element method, the matrix arose from the mass times acceleration term in Newton’s second law of motion. Making the mass matrix diagonal by, e.g., numerical integration, as demonstrated above, is a widely used technique and is called mass lumping. In time-dependent problems it can sometimes enhance the numerical accuracy and computational efficiency of the finite element method. However, there are also examples where mass lumping destroys accuracy.