Abstract
In this work we propose and analyze a numerical method for electrical impedance tomography to recover a piecewise constant conductivity from boundary voltage measurements. It is based on standard Tikhonov regularization with a Modica–Mortola penalty functional and adaptive mesh refinement using suitable a posteriori error estimators of residual type that involve the state, adjoint and variational inequality in the necessary optimality condition and a separate marking strategy. We prove the convergence of the adaptive algorithm in the following sense: the sequence of discrete solutions contains a subsequence convergent to a solution of the continuous necessary optimality system. Several numerical examples are presented to illustrate the convergence behavior of the algorithm.
Export citation and abstract BibTeX RIS
Original content from this work may be used under the terms of the Creative Commons Attribution 3.0 licence. Any further distribution of this work must maintain attribution to the author(s) and the title of the work, journal citation and DOI.
1. Introduction
Electrical impedance tomography (EIT) aims at recovering the electrical conductivity distribution of an object from voltage measurements on the boundary. It has attracted much interest in medical imaging, geophysical prospecting, nondestructive evaluation and pneumatic oil pipeline conveying, etc. A large number of reconstruction algorithms have been proposed; see, e.g. [3, 15, 18, 21, 24, 26, 28, 30, 31, 33–37, 39, 40, 51, 54, 57] for a rather incomplete list. One prominent idea underpinning many imaging algorithms is regularization, especially Tikhonov regularization [29]. In practice, they are customarily implemented using the continuous piecewise linear finite element method (FEM) on quasi-uniform meshes, due to its flexibility in handling spatially variable coefficients and general domain geometry. The convergence analysis of finite element approximations was carried out in [22, 27, 47].
In several practical applications, the physical process is accurately described by the complete electrode model (CEM) [14, 50]. It employs nonstandard boundary conditions to capture characteristics of the experiment. In particular, around the electrode edges, the boundary condition changes from the Neumann to Robin type, which, according to classical elliptic regularity theory [23], induces weak solution singularity around the electrode edges; see, e.g. [45] for an early study. Further, the low regularity of the sought-after conductivity distribution, especially that enforced by a nonsmooth penalty, e.g. total variation, can also induce weak interior singularities of the solution. Thus, a (quasi-)uniform triangulation of the domain can be inefficient for resolving these singularities, and the discretization errors around electrode edges and internal interfaces can potentially compromise the reconstruction accuracy. These observations motivate the use of an adaptive strategy to achieve the desired accuracy in order to enhance the overall computational efficiency.
For direct problems, the mathematical theory of adaptive FEM (AFEM), including a posteriori error estimation, convergence and computational complexity, has advanced greatly [1, 12, 44, 53]. A common AFEM consists of the following successive loops:
The module ESTIMATE employs the given problem data and computed solutions to provide computable quantities on the local errors, and distinguishes different adaptive algorithms.
In this work, we develop an adaptive EIT reconstruction algorithm with a piecewise constant conductivity. In practice, the piecewise constant nature is commonly enforced by a total variation (TV) penalty. However, it is challenging for AFEM treatment (see e.g. [5] for image denoising). Thus, we take an indirect approach based on a Modica–Mortola-type functional:
where the constant is small and is the double-well potential, i.e.
with being two known values that the conductivity can take. The functional converges to the TV semi-norm [2, 42, 43]. The corresponding regularized least-squares formulation reads
where is a regularization parameter; see section 2 for further details. In this work, we propose a posteriori error estimators and an adaptive reconstruction algorithm of the form (1.1) for (1.3) based on a separate marking using three error indicators in the module MARK; see algorithm 1. Further, we give a convergence analysis of the algorithm, in the sense that the sequence of state, adjoint and conductivity generated by the adaptive algorithm contains a convergent subsequence to a solution of the necessary optimality condition. The technical proof consists of two steps: step 1 shows the subsequential convergence to a solution of a limiting problem, and step 2 proves that the solution of the limiting problem satisfies the necessary optimality condition. The main technical challenges in the convergence analysis include the nonlinearity of the forward model, the nonconvexity of the double-well potential and properly treating the variational inequality. The latter two are overcome by pointwise convergence of discrete minimizers and Lebesgue's dominated convergence theorem, and AFEM analysis techniques for elliptic obstacle problems, respectively. The adaptive algorithm and its convergence analysis are the main contributions of this work.
Last, we situate this work in the existing literature. In recent years, several adaptive techniques, including AFEM, have been applied to the numerical resolution of inverse problems. In a series of works [6–9], Beilina et al studied the AFEM in a dual-weighted residual framework for parameter identification. Feng et al [20] proposed a residual-based estimator for the state, adjoint and control by assuming convexity of the cost functional and high regularity on the control. Li et al [38] derived a posteriori error estimators for recovering the flux and proved their reliability; see [55] for a plain convergence analysis. Clason et al [17] studied functional a posteriori estimators for convex regularized formulations. Recently, Jin et al [32] proposed a first AFEM for the Tikhonov functional for EIT with an penalty, and also provided a convergence analysis. This work extends the approach in [32] to the case of a piecewise constant conductivity. There are a number of major differences between this work and [32]. First, the penalty in [32] facilitates deriving the a posteriori estimator on the conductivity , by completing the squares and suitable approximation argument, which is not directly available for the Mordica–Mortola functional . Second, we develop a sharper error indicator associated with the crucial variational inequality than that in [32], by a novel constraint-preserving interpolation operator [13]; see the proof of theorem 5.5, which represents the main technical novelty of this work. Third, algorithm 1 employs a separate marking for the estimators, instead of the collective marking in [32], which automatically takes care of different scalings of the estimators.
The rest of this paper is organized as follows. In section 2, we introduce the CEM and the regularized least-squares formulation. In section 3, we give the AFEM algorithm. In section 4, we present extensive numerical results to illustrate its convergence and efficiency. In section 5, we present the lengthy technical convergence analysis. Throughout, and denote the inner product on the Euclidean space and , respectively, by the Euclidean norm, and occasionally abuse for the duality pairing between the Hilbert space and its dual space. The superscript denotes the transpose of a vector. The notation c denotes a generic constant, which may differ at each occurrence, but it is always independent of the mesh size and other quantities of interest.
2. Regularized approach
This part describes the regularized approach for recovering piecewise constant conductivities.
2.1. Complete electrode model (CEM)
Let be an open bounded domain in with a polyhedral boundary . We denote the set of electrodes by , which are line segments/planar surfaces on and satisfy if . The applied current on electrode el is denoted by Il, and the vector satisfies , i.e. . The electrode voltage is normalized, i.e. . Then the CEM reads: given the conductivity , positive contact impedances and input current , find such that [14, 50]
The inverse problem is to recover the conductivity from a noisy version of the electrode voltage (for the exact conductivity ) corresponding to one or multiple input currents.
Below the conductivity is assumed to be piecewise constant, i.e. in the admissible set
where the constants are known, is an unknown open set with a Lipschitz boundary and denotes its characteristic function. We denote by the space with its norm given by
A convenient equivalent norm on the space is given below.
Lemma 2.1. On the space , the norm is equivalent to the norm defined by
The weak formulation of the model (2.1) reads [50]: find such that
where the trilinear form on is defined by
where denotes the inner product. For any , and , the existence and uniqueness of a solution to (2.2) follows from lemma 2.1 and the Lax–Milgram theorem.
2.2. Regularized reconstruction
For numerical reconstruction with a piecewise constant conductivity, the TV penalty is popular. The conductivity is assumed to be in the space of bounded variation [4, 19], i.e.
equipped with the norm , where
Below we discuss only one dataset, since the case of multiple datasets is similar. Then Tikhonov regularization leads to the following minimization problem:
The scalar is known as a regularization parameter. It has at least one minimizer [22, 46].
Since is piecewise constant, by Lebesgue decomposition theorem [4], the TV term in (2.3) reduces to , where is the jump set, denotes the jump across and refers to the -dimensional Hausdorff measure. The numerical approximation of (2.3) requires simultaneously treating two sets of different Hausdorff dimensions (i.e. and ), which is very challenging. Thus, we replace the TV term in (2.3) by a Modica–Mortola-type functional [43]:
where is a small positive constant controlling the width of the transition interface, and is the double-well potential given in (1.2). The functional was first proposed to model the phase transition of two immiscible fluids in [11]. It is connected with the TV semi-norm as follows [2, 42, 43]; see [10] for an introduction to -convergence.
Then converges to in as . Let and be given sequences such that and is bounded. Then is precompact in .
The proposed EIT reconstruction method reads
where , and the admissible set is defined as
Now we recall a useful continuity result of the forward map [22, lemma 2.2], which gives the continuity of the fidelity term in the functional . See also [18, 30] for related continuity results.
Lemma 2.2. Let satisfy in . Then
Lemma 2.2 implies that are continuous perturbations of in . Then the stability of -convergence [2, proposition 1(ii)] [10, remark 1.4] and theorem 2.1 indicate that converges to with respect to , and can (approximately) recover piecewise constant conductivities. Next we show the existence of a minimizer to .
Theorem 2.2. For each , there exists at least one minimizer to problem (2.4).
Proof. Since is non-negative, there exists a minimizing sequence such that . Thus, , which, along with , yields . Since is closed and convex, there exists a subsequence, relabeled as , and some such that
Since , is uniformly bounded in and converges to almost everywhere in . By Lebesgue's dominated convergence theorem [19, p 28, theorem 1.19], By lemma 2.2 and the weak lower semi-continuity of the semi-norm, we obtain
Thus is a global minimizer of the functional . □
To obtain the necessary optimality system of (2.4), we use the standard adjoint technique. The adjoint problem for (2.2) reads: find such that
By straightforward computation, the Gâteaux derivative of the functional at in the direction is given by
Then the minimizer to problem (2.4) and the respective state and adjoint satisfy the following necessary optimality system:
where the variational inequality in the last line is due to the box constraint in the admissible set . The optimality system (2.8) forms the basis of the adaptive algorithm and its convergence analysis.
3. Adaptive algorithm
Now we develop an AFEM for problem (2.4). Let be a shape-regular triangulation of into simplicial elements, each intersecting at most one electrode surface el, and be the set of all possible conforming triangulations of obtained from by the successive use of bisection. Then is uniformly shape-regular, i.e. the shape-regularity of any mesh is bounded by a constant depending only on [44, 52]. Over any , we define a continuous piecewise linear finite element space
where P1(T) consists of all linear functions on T. The space is used for approximating the state u and adjoint p , and the discrete admissible set for the conductivity is given by
Given , the discrete analogue of problem (2.2) is to find such that
Then we approximate problem (2.4) by minimizing the following functional over :
Then, similar to theorem 2.2, there exists at least one minimizer to (3.2), and the minimizer and the related state and adjoint satisfy
Further, and depend continuously on the problem data, i.e.
where the constant c can be made independent of and .
To describe the error estimators, we first recall some useful notation. The collection of all faces (respectively all interior faces) in is denoted by (respectively ) and its restriction to the electrode and by and , respectively. A face/edge F has a fixed normal unit vector in with on . The diameter of any and is denoted by and , respectively. For the solution to problem (3.3), we define two element residuals for each element and two face residuals for each face by
where denotes the jump across interior face F. Then for any element , we define the following three error estimators
with . The estimator is identical to the standard residual error indicator for the direct problem: find such that
It differs from the direct problem in (2.8) by replacing the conductivity with instead, and is a perturbation of the latter case. The perturbation is vanishingly small in the event of the conjectured (subsequential) convergence . The estimator admits a similar interpretation. These two estimators are essentially identical to that for the penalty in [32], and we refer to [32, section 3.3] for a detailed heuristic derivation. The estimator is related to the variational inequality in the necessary optimality condition (2.8), and roughly provides a quantitative measure of how well it is satisfied. The estimator (including the exponent q) is motivated by the convergence analysis; see the proof of theorem 5.5 and remark 5.2 below. It represents the main new ingredient for problem (2.4), and differs from that for the penalty in [32].
Remark 3.1. The estimator improves that in [32], i.e.
in terms of the exponents on hT and hF. This improvement is achieved by a novel constraint-preserving interpolation operator defined in (5.13) below.
Now we can formulate an adaptive algorithm for (2.4); see algorithm 1. Below we indicate the dependence on the mesh by the subscript k, e.g. for .
Algorithm 1. AFEM for EIT with a piecewise constant conductivity.
1: Specify an initial mesh , and set the maximum number K of refinements. |
2: for do |
3: (SOLVE) Solve problem (3.1) and (3.2) over for and (3.3) for . |
4: (ESTIMATE) Compute error indicators , and . |
5: (MARK) Mark three subsets () such that each contains at least one element |
() with the largest error indicator: |
Then . |
6: (REFINE) Refine each element T in by bisection to get . |
7: Check the stopping criterion. |
8: end for |
9: Output . |
The MARK module selects a collection of elements in the mesh . The condition (3.5) covers several commonly used marking strategies, e.g. maximum, equidistribution, modified equidistribution and Dörfler's strategy [49, pp 962]. Compared with a collective marking in AFEM in [32], algorithm 1 employs a separate marking to select more elements for refinement in each loop, which leads to fewer iterations of the adaptive process. The error estimators may also be used for coarsening, which is relevant if the recovered inclusions change dramatically during the iteration. However, the convergence analysis below does not carry over to coarsening, and it will not be further explored.
Last, we give the main theoretical result: for each fixed , the sequence of discrete solutions generated by algorithm 1 contains a subsequence converging in to a solution of system (2.8). The proof is lengthy and technical, and thus deferred to section 5.
Theorem 3.1. The sequence of discrete solutions by algorithm 1 contains a subsequence convergent to a solution of system (2.8):
4. Numerical experiments and discussions
Now we present numerical results to illustrate algorithm 1 on a square domain . There are sixteen electrodes (with L = 16) evenly distributed along , each of length 1/4. The contact impedances are all set to unit. We take ten sinusoidal input currents, and for each voltage , generate the noisy data by
where is the (relative) noise level, and follow the standard normal distribution. Note that refers to a relatively high noise level for EIT. The exact data is computed using a much finer uniform mesh, to avoid the most obvious form of 'inverse crime'.
In the experiments, we fix K (the number of refinements) at 15, q (exponent in ) at 2, and (the functional ) at . The marking strategy (3.5) in the module MARK selects a minimal refinement set such that
with a threshold . The refinement is performed with one popular refinement strategy, i.e. newest vertex bisection [41]. Specifically, it connects the midpoint xT, as a newest vertex, of a reference edge F of an element to the opposite node of F, and employs two edges opposite to the midpoint xT as reference edges of the two newly created triangles in . Problems (3.1) and (3.2) are solved by a Newton-type method; see appendix A for the details. The conductivity on is initialized to , and then for , (defined on ) is interpolated to to warm-start the optimization. The regularization parameter in (2.4) is determined in a trial-and-error manner. All computations are performed using MATLAB 2018a on a personal laptop with 8.00 GB RAM and 2.5 GHz CPU.
The first set of examples is concerned with two inclusions.
Example 1. The background conductivity .
- (i)The true conductivity is given by , where B1 and B2 denote two circles centered at and , respectively, both with a radius 0.3.
- (ii)The true conductivity is given by , i.e. two Gaussian bumps centered at and .
- (iii)The true conductivity is given by , where B1 and B2 denote two circles centered at and , respectively, both with a radius of 0.3.
The numerical results for example 1(i) with and are shown in figures 1–5, where d.o.f. denotes the degree of freedom of the mesh. It is observed from figure 1 that with both uniform and adaptive refinements, the final recoveries have comparable accuracy and capture the inclusion locations well.
Download figure:
Standard image High-resolution imageDownload figure:
Standard image High-resolution imageDownload figure:
Standard image High-resolution imageDownload figure:
Standard image High-resolution imageNext we examine the adaptive refinement process more closely. In figures 2 and 3, we show the meshes during the iteration and the corresponding recoveries for example 1(i) at two noise levels and , respectively. On the coarse mesh , the recovery has very large errors and can only identify one component and thus fails to correctly identify the number of inclusions, due to the severe under-resolution of both state and conductivity. Nonetheless, algorithm 1 can correctly recover the two components with reasonable accuracy after several adaptive loops, and accordingly, the support of the recovery is gradually refined with its accuracy improving steadily. In particular, the inclusion locations stabilize after several loops, and thus coarsening of the mesh seems unnecessary. Throughout, the refinement occurs mainly in the regions around the electrode edges and internal interface, which is clearly observed for both noise levels. This is attributed to the separable marking strategy, which allows detecting different sources of singularities simultaneously. In figure 4, we display the evolution of the error indicators for example 1(i) with . The estimators play different roles: and indicate the electrode edges during first iterations and then also the internal interface, whereas throughout concentrates on the internal interface. Thus, and are most effective for resolving the state and adjoint, whereas is effective for detecting internal jumps of the conductivity. The magnitude of is much smaller than , since the boundary data for the adjoint are much smaller than the input current I for the state. Thus, a simple collective marking strategy (i.e. ) may miss the correct singularity, due to their drastically different scalings. In contrast, the separate marking in (3.5) can take care of the scaling automatically.
In figure 5, we plot the and errors of the recoveries versus d.o.f. N, where the recovery on the corresponding finest mesh is taken as the reference (since the recoveries by the adaptive and uniform refinements are slightly different; see figure 1). Due to the discontinuity of the sought-after conductivity, the norm is especially suitable for measuring the convergence. The convergence of the algorithm is clearly observed for both adaptive and uniform refinements. Further, with a fixed d.o.f., AFEM gives more accurate results than the uniform one in both error metrics. These observations show the computational efficiency of the adaptive algorithm.
Examples 1(ii) and (iii) are variations of example 1(i), and the results are presented in figures 6–9. The proposed approach assumes a piecewise constant conductivity with known lower and upper bounds. Example 1(ii) does not fulfill the assumption, since the true conductivity is not piecewise constant. Thus the algorithm can only produce a piecewise constant approximation to the exact one. Nonetheless, the inclusion support is reasonably identified. When the noise level increases from to , the reconstruction accuracy deteriorates significantly; see figure 6. Example 1(iii) involves high-contrast inclusions, which are well known to be numerically more challenging. This is clearly observed in figure 8, where the recovery accuracy is inferior, especially for the noise level . However, the adaptive refinement procedure works well similarly to the preceding examples: the refinement occurs mainly around the electrode edges and inclusion interface; see figures 7 and 9 for the details.
Download figure:
Standard image High-resolution imageDownload figure:
Standard image High-resolution imageDownload figure:
Standard image High-resolution imageDownload figure:
Standard image High-resolution imageNow we consider one more challenging example with four inclusions.
Example 2. The true conductivity is given by , with the circles Bi centered at and , respectively, all with a radius 0.2, and the background conductivity .
The numerical results for example 2 are given in figures 10–12. The results are in excellent agreement with the observations from example 1: the algorithm converges steadily as the adaptive iteration proceeds, and with a low noise level, it can accurately recover all four inclusions, showing clearly the efficiency of the adaptive approach. The refinement is mainly around the electrode edge and interval interface.
Download figure:
Standard image High-resolution imageDownload figure:
Standard image High-resolution imageDownload figure:
Standard image High-resolution image5. Proof of theorem 3.1
The lengthy and technical proof is divided into two steps: step 1 shows the convergence to an auxiliary minimization problem over a limiting admissible set in section 5.1, and step 2 shows that the solution of the auxiliary problem satisfies the necessary optimality system (2.8) in section 5.2. The overall proof strategy is similar to [32], and hence we omit the relevant arguments whenever possible.
5.1. Auxiliary convergence
Since the two sequences and generated by algorithm 1 are nested, we may define
Clearly is a closed subspace of . For the set , we have the following result [32, lemma 4.1].
Lemma 5.1. is a closed convex subset of .
Over the limiting set , we define an auxiliary limiting minimization problem:
where satisfies
By lemma 2.1 and the Lax–Milgram theorem, problem (5.2) is well posed for any fixed . The next result gives the existence of a minimizer to (5.1) and (5.2).
Theorem 5.1. There exists at least one minimizer to problem (5.1)–(5.2).
Proof. Let be the sequence of discrete solutions given by algorithm 1. Since for all k, by (3.4), , and thus is uniformly bounded in . By lemma 5.1 and Sobolev embedding, there exists a subsequence, denoted by , and some such that
Next we introduce a discrete analogue of problem (5.2) with : find such that
By lemma 2.1, Cea's lemma and the construction of the space , the solution of (5.2) with satisfies
Taking the test function in the first line of (3.3) and (5.4) and then applying the Cauchy–Schwarz inequality lead to
In view of (5.5), pointwise convergence in (5.3) and Lebesgue's dominated convergence theorem,
This and lemma 2.1 imply Then, (5.5) and the triangle inequality imply
Meanwhile, repeating the argument of theorem 2.2 gives
Next we apply a density argument. For any , by the construction of the space , there exists a sequence such that in . Repeating the preceding argument gives and . Now (5.6), the weak lower semicontinuity of the -norm, (5.7) and the minimizing property of to over the set imply
Since , is a minimizer of over . □
Further, we have the following auxiliary convergence.
Theorem 5.2. The sequence of discrete solutions to problem (3.2) contains a subsequence convergent to a minimizer to problem (5.1)–(5.2):
Proof. The convergence of was already proved in theorem 5.1. Taking in (5.8) gives . By (5.6) and (5.7), we have . Thus, the sequence converges to in . □
Next we consider the convergence of the sequence . With a minimizer to problem (5.1), we define a limiting adjoint problem: find such that
By lemma 2.1 and Lax–Milgram theorem, (5.9) is uniquely solvable. We have the following convergence result for . The proof is identical to [32, theorem 4.5], and hence omitted.
Theorem 5.3. Under the condition of theorem 5.2, the subsequence of adjoint solutions generated by algorithm 1 converges to the solution of problem (5.9):
5.2. Proof of theorem 3.1
Theorem 3.1 follows directly by combining theorems 5.2 and 5.3 in section 5.1 and theorems 5.4 and 5.5 below. The proof in this part relies on the marking condition (3.5). First, we show that the limit solves the variational equations in (2.8).
Theorem 5.4. The solutions and to problems (5.1), (5.2) and (5.9) satisfy
Proof. The proof is identical to [32, lemma 4.8], using theorems 5.2 and 5.3, and hence we only give a brief sketch. By [32, lemma 3.5], for each with its face F (intersecting with el), there hold
where the notation DT is defined below. Then by the marking condition (3.5), [32, lemma 4.6] implies that for each convergent subsequence from theorems 5.2 and 5.3, there hold
Last, by [32, lemma 4.7] and theorems 5.2 and 5.3, the argument of [32, lemma 4.8] completes the proof. □
Remark 5.1. The argument of theorem 5.4 dates back to [49], and the main tools include the Galerkin orthogonality of the residual operator, the Lagrange and the Scott–Zhang interpolation operators [16, 48], the marking condition (3.5) and a density argument. Further, the error estimators and emerge in the proof and are then employed in the module ESTIMATE of algorithm 1.
Next we prove that the limit satisfies the variational inequality in (2.8). The proof relies crucially on a constraint-preserving interpolation operator. We denote by DT the union of elements in with a non-empty intersection with an element , and by the union of elements in sharing a common face/edge with . Let
The set consists of all elements not refined after the kth iteration, and all elements in are refined at least once after the kth iteration. Clearly, for l < k. We also define a mesh-size function almost everywhere
where Ti denotes the interior of an element , and Fi the relative interior of an edge . It has the following property [49, corollary 3.3]:
The next result gives the limiting behavior of the maximal error indicator .
Lemma 5.2. Let be the sequence of discrete solutions generated by algorithm 1. Then for each convergent subsequence , there holds
Proof. The inverse estimate and scaled trace theorem imply that for each (with its face F)
With the choice , combining these two estimates gives
where c depends on and in . Next, for the subsequence , let be the element with the largest error indicator . Since , (5.10) implies
By (5.11), Cauchy–Schwarz inequality and triangle inequality, there holds
By theorems 5.2 and 5.3, Lebesgue's dominated convergence theorem, the choice and Hölder inequality, we obtain and . Then the absolute continuity of the norm with respect to Lebesgue measure and (5.12) complete the proof. □
Due to a lack of Galerkin orthogonality for variational inequalities, we employ a local Lr-stable interpolation operator of Clément/Chen–Nochetto type. Let be the set of all interior nodes of and be the nodal basis functions in . For each , the support of is denoted by , i.e. the union of all elements in with a non-empty intersection with x. Then we define by
Clearly, if a.e. . The definition is adapted from [13] (for elliptic obstacle problems) by replacing the maximal ball centered at an interior node x by . satisfies the following properties; see appendix B for a proof.
Lemma 5.3. For any , there hold for all , any and any ,
Last we show that the limit satisfies the variational inequality in (2.8).
Theorem 5.5. The solutions and to problems (5.1), (5.2) and (5.9) satisfy
Proof. The proof is lengthy, and we break it into five steps.
Step i. Derive a preliminary variational inequality. We relabel the subsequence in theorems 5.2 and 5.3 as . Let Ik be the Lagrange interpolation operator on , and let and . For any , and let . Direct computation gives
where the last inequality is due to the variational inequality in (3.3) with .
Step ii. Bound the term . By elementwise integration by parts, Hölder inequality, the definition of the estimator and lemma 5.3 with (with being the conjugate exponent of q),
Thus, for any k > l, by (discrete) Hölder's inequality and the finite overlapping property of the patches DT, due to uniform shape regularity of the meshes , there holds
Since , by the pointwise convergence of in theorem 5.2 and Lebesgue's dominated convergence theorem, we deduce
Since , the sequence is uniformly bounded in . By theorems 5.2 and 5.3, the sequences , and are uniformly bounded in . Thus, (5.11) and (5.10), and Hölder inequality give
Then by the error estimate of Ik [16],
By (5.10), as . Since for k > l, (3.5) implies
By lemma 5.2, for any small , we can choose for some large fixed l1 such that whenever k > k1,
Consequently,
Step iii. Bound the term . For the term , elementwise integration and Hölder inequality yield
By the scaled trace theorem, local inverse estimate, -stability of in lemma 5.3, local quasi-uniformity and interpolation error estimate for Ik [16], we deduce that for k > l
Since , see (5.16), there holds
Now by repeating the argument for the term , we obtain
Step iv. Take limit in preliminary variational inequality. Using (5.15) and the -convergence of in Theorem 5.2, we have for each
Further, the uniform boundedness on in and the convergence of to in in theorem 5.3 yield
This and theorem 5.2 imply
In the splitting
the arguments for (5.20) directly yield
The boundedness on in , pointwise convergence of of theorem 5.2 and Lebesgue's dominated convergence theorem imply
Hence, there holds
Now by passing both sides of (5.14) to the limit and combining the estimates (5.17)–(5.21), we obtain
Step v. Density argument. By the density of in and the construction via a standard mollifier [19], for any there exists a sequence such that as . Thus, , , and , after possibly passing to a subsequence. The desired result follows from the preceding two estimates. □
Remark 5.2. The computable quantity emerges naturally from the proof, i.e. the upper bounds on and , which motivates its use as the a posteriori error estimator in algorithm 1.
Acknowledgments
The authors are grateful to an anonymous referee and the boarder member for the constructive comments, which have significantly improved the presentation of the paper. The work of Y Xu was partially supported by the National Natural Science Foundation of China (Grant No. 11201307), the Ministry of the Education of China through the Specialized Research Fund for the Doctoral Program of Higher Education (Grant No. 20123127120001) and the Natural Science Foundation of Shanghai (Grant No. 17ZR1420800).
Appendix A. The solution of the variational inequality
Now we describe an iterative method for minimizing the energy functional
Let . Then one linearized approximation reads (with )
Upon substituting the approximation for and linearizing the forward map , we obtain the following surrogate energy functional (with being the increment and )
The treatment of the double-well potential term is in the spirit of the classical majorization–minimization algorithm in the following sense (see [56] for a detailed derivation)
This algorithm is known to have excellent numerical stability. Upon ignoring the box constraint on the conductivity , problem (A.1) is to find such that
This equation can be solved by an iterative method for the update (with the box constraint treated by a projection step). Note that and can be implemented in a matrix-free manner using the standard adjoint technique. In our experiment, we employ the conjugate gradient method to solve the resulting linear systems, preconditioned by the sparse matrix corresponding to .
Appendix B.: Proof of lemma 5.3
The proof follows that in [13, 25]. By Hölder inequality and for each node :
The desired Lr-stability follows from the estimate , by the local quasi-uniformity of the mesh. In view of the definition (5.13), for any . By local inverse estimate, the Lr-stability of , standard interpolation error estimate [16] and local quasi-uniformity,
Similarly,
By the scaled trace theorem, for any , there holds