![MathJax Logo](/templates/jsp/_style2/_tandf/pb2/images/math-jax.gif)
Abstract
For the investigation of two-component aerosols, one needs to know the refractive indices of the two aerosol components. One problem is that they depend on temperature and pressure, so one needs for their determination a robust measurement instrument such as the FASP device, which can cope with rigid environmental conditions. In this article, we show that the FASP device is capable of measuring the needed refractive indices, if monodisperse aerosols of the pure components are provided. We determine the particle radii of the monodisperse aerosols needed for this task and investigate how accurate the measurements have to be in order to retrieve refractive indices in a sufficient quality, such that they are suitable for investigations of two-component aerosols.
1. Set-up of the experiment
This paper provides an algorithm for the reconstruction of refractive indices from spectral measurements of monodisperse aerosols. The experiments we are conducting are similar to the experiments presented in [Citation1] with the difference that we are using air as surrounding medium for the aerosol particles and that temperature and pressure may approach and 8 bar, respectively. For these rigid conditions, reliable databases for refractive indices do not exist up to now. These refractive index databases are needed for the measurement of particle size distributions of polydisperse aerosols using the FASP.
As outlined in [Citation2], the FASP measures light intensities and
having passed a long and a short measurement path length
and
, respectively. The evaluations of the FASP measurements are based on the relation
(1.1)
(1.1)
where is the so-called kernel function, l is the wavelength of the incident light, r is the radius of the spherical scattering particle and
and
are the refractive indices of the surrounding medium and the particle material depending on the wavelength l. The function
is the Mie extinction efficiency from [Citation3]. The function n(r) is the size distribution of the scattering particels. The right-hand side e(l) in (Equation1.1
(1.1)
(1.1) ) is denoted as the spectral extinction.
Now if n(r) is the size distribution of a monodisperse aerosol, where all particles possess the same radius , it is given by
, where n is the total number of particles and
is a Dirac delta distribution truncated on the positive half-axis. Inserting this into (Equation1.1
(1.1)
(1.1) ) gives
(1.2)
(1.2)
hence the Mie extinction efficiency is measured directly at the radius .
The Mie extinction efficiency is given as an infinite series, i.e.
The computation of the coefficient functions will be discussed in Section 2.
It is clear that in practical computations can only be approximated by a truncated series, because only the computation of a finite number of the
’s is practically feasible.
We now fix a wavelength l. The complex refractive index for the wavelength l is reconstructed from FASP measurements of several monodisperse aerosols with particle radii
, ...,
. Let
, ...,
denote the measured spectral extinctions e(l) corresponding to the particle radii
, ...,
. We assume that they are contaminated by additive Gaussian noise, i.e.
with
for
. Furthermore, we assume that the standard deviations
can be estimated from measurements sufficiently accurately, such that we can regard them as known. We have
where is the number of particles having the same radius
. Then, a reconstrution of
is obtained from the set of solutions M(l) of the nonlinear regression problem
(1.3)
(1.3)
Note that M(l) contains in general more than one solution, especially when is perturbed by measurement noise. We discuss nonlinear regression problems with truncated series expansions such as (Equation1.3
(1.3)
(1.3) ) in Section 4.
For solving (Equation1.3(1.3)
(1.3) ), we use a global optimization strategy presented in Section 5 to generate reasonable candidates for start values for a local solver for a regularized version of (Equation1.3
(1.3)
(1.3) ). Section 8 provides a selection method to find a unique start value out of the candidates. In order to apply a gradient-based local solver, we must know the derivatives of the Mie extinction efficiency series, which are discussed in Section 3.
2. Mie theory
We recapitulate Mie theory in an absorbing medium as presented in [Citation3]. Our first step is to introduce the complex-valued Riccati–Bessel functions and
given by
(2.1)
(2.1)
and(2.2)
(2.2)
with the Bessel functions and
of order
of first and second kinds. We define the size parameter
. Then, we set
and
. Here and in the following, we omit the wavelength dependence of
and
for better readability. We introduce the notation
and
.
We introduce the so-called Mie coefficients:(2.3)
(2.3)
With the Mie coefficients, we can express the coefficient functions(2.4)
(2.4)
and(2.5)
(2.5)
which finally occur in the series expansion(2.6)
(2.6)
Here the quantity I(r, l) is the average incident intensity of light with wavelength l for a spherical particle with radius r and c denotes the speed of light in vacuum. The function I(r, l) is given by(2.7)
(2.7)
Obviously we cannot evaluate (Equation2.6(2.6)
(2.6) ) exactly, because we cannot compute an infinite sum due to limited processing resources. Therefore, we have to truncate this series expansion. In [Citation4], a commonly used truncation index
is presented, which is given by
(2.8)
(2.8)
3. Derivatives of the truncated Mie efficiency series
The Bessel functions and
for an arbitrary weight
fulfil the recurrence relations
(3.1)
(3.1)
cf. [Citation5]. For the Bessel functions occurring in the Riccati–Bessel functions, and
follow from this with the weight
that
(3.2)
(3.2)
(3.3)
(3.3)
for .
We apply (Equation3.1(3.1)
(3.1) ) a second time to get
, which yields
For an arbitrary weight , we have the recurrence relation
see [Citation5], and we use it to eliminate the term in the expression for
. Then also
cancels out, such that we obtain the representation
(3.4)
(3.4)
only involving .
We recapitulate the Cauchy–Riemann equations in its complex form. For a holomorphic function with
holds
(3.5)
(3.5)
From this follows(3.6)
(3.6)
Using the latter relations, we can deduce using
and analogously
The Bessel function values
already computed for a function evaluation of the truncated Mie extinction efficiency can be reused for their derivatives.
Now everything is prepared to differentiate the squared magnitude of the Mie coefficient with respect to
and
. It is sufficient only to discuss
in the following because the differentiation of
,
and
works analogously.
First we write the squared norm of the Mie coefficient as
, which gives
We write
which yields
Furthermore follow from (Equation3.5(3.5)
(3.5) ), the relations
Although is not holomorphic with respect to
, we can still compute the partial derivatives
and
. We obtain using (Equation3.6
(3.6)
(3.6) ) the relations
This completes the computations of and
.
For a holomorphic function , we can easily deduce from (Equation3.6
(3.6)
(3.6) ) that
This gives with respect to (Equation3.5(3.5)
(3.5) )
Analogously we obtain
The derivatives of are much easier to compute, since the dependence on
and
lies only in
and
here. So we can deduce
Analogously we get
4. Nonlinear regression using truncated series expansions
We wish to reconstruct the refractive indices of a particle material from spectral measurements by solving a nonlinear regression problem of the form(4.1)
(4.1)
where is a finite truncation index and
. Remember that N represents the number of particle radii
of the different monodisperse aerosols we are investigating. We still assume that for each radius
the standard deviations
are determined well enough from a set of experiments, such that they can be regarded as known. We have to confine ourselves to a finite truncation index t because it is practically not feasible to compute all coefficient functions
for
. Throughout this paper, we assume that the feasible set
is compact.
We define the functions and
by
We set with
. Then the observed probability density is given by
with the covariance matrix . We know a priori that the vector
specifying our model
lies within the set
. This knowledge can be expressed with the prior probability density
where is the indicator function of
. Now
is the set of MAP-estimators of the posterior probability density, i.e.
(4.2)
(4.2)
We carry out all the following investigations under the next assumption on the covariance matrix:
Assumption 4.1:
The covariance matrix has the simple form
where is an arbitrary but fixed noise level and
, ...,
are fixed.
To simplify notations, we introduce the two functions and
depending on the truncation index t and defined by
In the following, we will investigate how an element of the set
depends on the truncation index t. We change to a continuous truncation index here, i.e. we change from now on from (Equation4.1
(4.1)
(4.1) ) to the new regression problem
(4.3)
(4.3)
where the truncation index is allowed to be non-integer.
As a preparation we prove the following technical lemma, which will form the basis of our continuity and convergence results.
Lemma 4.2:
Let the twice continuously differentiable function have a strict local minimum
inside a compact set
. Let the function
have the property
for all
and let
be twice continuously differentiable with respect to
and continuous in
. Furthermore, we assume that the local minima
of
are strict for any
. Then there exists a sequence of local minima
of
with
.
The strategy of the proof is to construct for given a neighbourhood of
which must contain a local minimizer
of the perturbed function
. By sending
to 0, this neighbourhood shrinks down to the local minimum
itself, thus yielding the convergence of
to
. To have this neighbourhood shrink down to
, it is crucially important that
must be a strict local minimum.
We define . From
for all
follows
. Let us now introduce the function
. Obviously,
is also a local minimum of
, so for
sufficiently small there exists a neighbourhood
of
with
In particular, we have
Let us assume that there exists an with
Then implies
, hence
by definition of
, contradiction. Therefore we conclude
(4.4)
(4.4)
Since is continuous and
is compact for
small enough, there exists an
with
Let us assume . Then by definition of
, we get in particular
i.e. , contradiction. It follows
(4.5)
(4.5)
where the latter follows with a proof by contradiction as well.
If it happens to hold that , then we also have
. Otherwise we have
and then (Equation4.4
(4.4)
(4.4) ) implies that
cannot lie on
, thus it must lie within the interior of
. So in any case (Equation4.5
(4.5)
(4.5) ) gives that
must contain a local minimizer
of
.
Now gives
. The existence of the last limit is guaranteed by the fact that
is strict and the claim is proved.
Proposition 4.3:
Let all coefficient functions be twice continuously differentiable and bounded on
. We assume that each local minimum
of the right-hand side function
in (Equation4.3
(4.3)
(4.3) ) is strict and lies in the interior of
. Then each local minimum depends continuously on the truncation index t.
To prove the claim, one could be tempted to apply the implicit function theorem on the equation with
. This would give that the local minima are parameterized by a function
with the property
, where s is from an environment U(t) of t. The problem with this approach is that it requires continuous differentiability of
in the truncation parameter s. Thus the continuous truncation we are using would need more complicated methods such as spline interpolation of the partial sums, which would increase the overall computational effort.
Therefore, we use in the following a more direct approach to prove the claim. Let be arbitrary. First we consider an integer truncation index
, i.e. we have
. Now for
small enough, we get
and
. This gives
As next step we turn to an noninteger truncation index t. In this case, we can always select small enough such that
and
, respectively hold. This yields
Now we introduce the function
For this yields
Therefore, we obtain both for and
a decomposition of the form
and
respectively, where the function
is appropriately selected according to above findings. We can readily check
for all
from the boundedness of the
’s. Then the result follows from Lemma 4.2.
Corollary 4.4:
Let and
be truncation indices with
. Let
be a local minimizer of (Equation4.1
(4.1)
(4.1) ). Let
and define
. Then beginning at
one can successively find local minimizers
for the truncation index
using numerical continuation, see [Citation6]. Here for
the minimizer
is used as a start vector to compute the next minimizer
. The next parameter
has to be sufficiently close to
, such that the start vector
still lies within the domain of convergence for Newton’s method.
We use Corollary 4.4 to compute for increasing truncation index t in a stable way. If we would keep t as integer and increase it in integer steps, we might leave the domain of convergence in the continuation method. Therefore, we increase them using a smaller step width.
In the following, we investigate how well the minimizers of the noise-contaminated regression problem (Equation4.3
(4.3)
(4.3) ) with truncated series expansions approximate the minimizers
of the noise-free and untruncated problem
(4.6)
(4.6)
Proposition 4.5:
Let the noise vector fulfil
and let the functions
and
be bounded on
. Assume
for all
. Then for any strict minimizer
of the right-hand side function of (Equation4.6
(4.6)
(4.6) ) in the interior of
exist minimizers
of (Equation4.3
(4.3)
(4.3) ) with
. Here we also assume the
’s to be strict for all
.
With the notation introduced before, we can write
From the decomposition we obtain
Then a further decomposition of the first term on the right-hand side yields
From the limit , the limit
and the boundedness of
follows
for arbitrary but fixed
.
Then the existence of the ’s follows from Lemma 4.2.
At last we study how the minimizers of (Equation4.3
(4.3)
(4.3) ) behave for
. We begin with a preparing corollary.
Corollary 4.6:
Let the assumptions of Proposition 4.5 hold. Then we have for any local minimizer of (Equation4.3
(4.3)
(4.3) ) approximating a local minimizer
of (Equation4.6
(4.6)
(4.6) ) for
with
that
The assumptions of Proposition 4.5 give
We have by continuity of that
. Then
gives the desired result.
Proposition 4.7:
Let the assumptions of Proposition 4.5 hold. Assume that the local minimizers of (Equation4.6
(4.6)
(4.6) ) with
form a discrete set
. Then the set
consisting of the limits
of local minimizers
of (Equation4.3
(4.3)
(4.3) ) with
coincides with
and there exists a noise level
such that all minimizers
approximating
are isolated for all
.
On the one hand from Proposition 4.5 we know that there exists a sequence of minimizers of (Equation4.3
(4.3)
(4.3) ) with
. Then
and Corollary 4.6 give
, which implies
.
On the other hand holds for with
that
which implies . In particular, this means by continuity of
that the vector
must be a local minimizer of
. Thus we have also shown
.
In the following, we number all elements of with the index k, i.e. we write
for
. Similarly we number all minimizers
with
approximating the
’s with
, i.e.
for
. Define
Since , we can find an error levels
such that
which holds for all for each k. Then for all
the
’s must have pairwise mutual distances greater than zero.
Now Proposition 4.7 gives that the number of local minima remains constant if the noise level
is small enough. It also yields that these local minima then form a set of separated continuous curves parametrized in
.
At last we wish to have an estimate of the convergence of the local minima of the truncated and noise contaminated problem to the local minima
of the noise-free and untruncated problem, which is useful for practical computations.
Proposition 4.8:
For the noise-level small enough, we can bound for any local minimum
the approximation error
with a positively weighted linear combination of the residual
, the truncation error
and the noise estimate
.
The first-order necessary conditions for a local minimum of at
and a local minimum of
at
yield in particular
Adding the last two inequalities yields(4.7)
(4.7)
Since is totally differentiable at
, we obtain
where fulfils
Since is positive definite, the expression
gives a norm on
. Because of the equivalence of all norms in
, there exists a constant
with
Since we can find a noise level
such that
for all
, where
is a constant with
. Then using
and (Equation4.7(4.7)
(4.7) ) we can estimate
i.e. this gives
(4.8)
(4.8)
for all .
We have
i.e. we can find constants and
such that
which gives the result.
Corollary 4.9:
Let the truncation indices depend on the vector of independent Gaussian random variables
with
such that
holds for all arbitrary but fixed
. Then we have for all minimizers
of (Equation4.6
(4.6)
(4.6) ) with
that for
sufficiently small there exist minimizers
of (Equation4.3
(4.3)
(4.3) ) with
Proposition 4.5 establishes the existence of the ’s. Corollary 4.6 gives
Proposition 4.7 gives that for sufficiently small there exists a constant
with
Set and
. Then the estimate for
in the proof of Proposition 4.7 gives
which proves the claim since Assumption 4.1 gives
.
The strategy for our retrieval algorithm is to start with an initial guess for the truncation index and try to find all local minima
. Then the truncation index is gradually increased and starting from
the continuation method is applied to find finally the local minima
. Motivated by Propositions 4.8 and 4.9 only those local minima are considered to be possible approximations to our sought-after refractive index, where the residual
and an estimate of the truncation error
are both reasonably small. The latter serves also as a stopping criterion for the continuation method
The initial guess has to be selected with care. On the one hand if it is to small, the model is to inaccurate and the retrieval of the sought-after local minima cannot be guaranteed. On the other hand if it is too big, computational effort is wasted, since too many Mie coefficient functions with almost vanishing magnitudes and thus essentially not changing the local minima are computed.
Remark 4.10:
We did not explicitly show the assumption that the local minima of (Equation4.3(4.3)
(4.3) ) are strict as demanded in Proposition 4.5. However we could always determine a set of strict local minima for any truncation index in our simulations.
5. The reconstruction algorithm
We now return to our regression problem (Equation1.3(1.3)
(1.3) ). For
we see that the measured extinctions normalized by the number of particles
with radius
, i.e. the quantity
, is Gaussian-distributed with mean
and standard deviation
. In the following we fix a wavelength l, i.e. we reconstruct the sought-after particle refractive index
wavelength by wavelength. In the following, the unit both for particle radii and light wavelengths is
m.
We make use of the function defined by
(5.1)
(5.1)
where is the particle radius grid and
the truncation index to be used. We allow non-integer truncation indices
as well, where the non-integer truncation is done like in Proposition 4.3. Here the expression
is a short notation of
from Section 1, where the sought-after refrative index
is identified with the vector
here, i.e.
. So its computation follows Section 2.
In the following, the refractive index search area is given by the rectangle , which means that we only consider refractive indices of particle materials whose real parts lie in the interval [0, 20] and its imaginary parts in the interval [0, 40]. This rather large search area makes the algorithm suitable for a wide range of aerosol materials.
In the first loop from lines 14–25, a search for local minima of the fit function defined in line 16 for the truncation index
is performed. The loop runs through all grid points
of the search grid defined in lines 5 and 6. If the Hessian of
at some grid point
is positive definite, this point might lie in the vicinity of a local minimum. The Hessian is computed exactly, where the second partial derivatives of the Mie extinction efficiency with respect to the real and imaginary part of the scattering material needed here are computed using the product rule approach from Section 3. So we use
as start point for a local solver in this case. In line 20, we only accept a new local minimum if it is sufficiently different from the local minima already found. Then it is stored in the container
. This simple global search strategy can find all local minima if the search grid is fine enough.
Remark 5.1:
Of course other well-established global search heuristics can be applied here as well. In test runs, we compared genetic algorithms with our sequential search strategy on the two-dimensional refractive index search area, but their computational effort and reliability remained the same or were even worse. In [Citation7], the technique of simulated annealing was used to retrieve aerosol refractive indices, which could be a promising alternative here. In our study, the measurement noise was so high, that a unique global minimum of our fit function could not be determined. Instead our focus lied on effectively finding all local minima with small values of the fit function and we regarded them all as possible approximations to a thought-after refractive index.
The second loop from lines 29–49 uses the local minima found in the first loop as start points for the continuation method following Proposition 4.3 and Corollary 4.4. We found that a step width of 0.1 is for our problem a well-balanced choice between too big step widths rendering the continuation method unstable and too small step widths making it computationally inefiicient. With the stopping criterion of the while-loop, it is approximately checked if the magnitude of the remainder term is small enough. Finally in line 44, it is checked if the residual is small enough. In our implementation, we did another run of lines 44–48 with
and
, respectively, if none of the reconstructions had a squared residual smaller than
for the previous
. This had to be done, because the parameter
has to be selected carefully in order to estimate the bound on
derived in the proof of Corollary 4.9 correctly.
6. Comparison of the Numerical Continuation Approach with Established Truncation Index Heuristics
As solution of the forward problem, we generated for a discrete set of wavelengths , ...,
unperturbed spectral extinctions normalized with the number of particles of the monodisperse aerosol by computing
with taken as the refractive indices of Ag,
O and CsI. Here we used the truncation index
(6.1)
(6.1)
introduced in [Citation4].
For particle size distribution reconstructions as outlined in [Citation2], we need particle refractive indices for five optical windows, see [Citation8], so the wavelength grid of interest consists of five ranges. These ranges are given by 8 linearly spaced wavelengths from 0.6–0.8 m, 8 from 1.1–1.3
m, 8 from 1.6–1.8
m, 16 from 2.1–2.5
m and 8 from 3.1–3.3
m, so we have in total
wavelengths.
For each of the 48 wavelengths, we generated noisy spectral extinctions by adding zero-mean Gaussian noise to
, i.e.
Here the standard deviations were taken to be of the original extinction values. We computed each mean
of the noisy spectral extinctions with a sample size of
.
6.1. Run Time Comparison
In the following, Algortihm Equation1(1.1)
(1.1) is referred to as method 1. On the same simulated spectral extinctions, we let Algorithm 1 run up to line 25, but with the difference that at each evaluation of
we directly took the trunction index from (Equation6.1
(6.1)
(6.1) ). We denote this approach with method 2. We now display the average run times of method 1 and method 2 for 10 sweeps through all 48 wavelenghs.
6.1.1. Results for Ag
6.1.2. Results for CsI
6.1.3. Results for ![](//:0)
O
6.2. Maximal relative deviations
For the 10 simulation runs, we list the maximal relative deviations
of the refractive index reconstructions from method 2 from
of method 1 for
. At each wavelength, multiple local minima can be detected by both methods. For the relative deviations, we always selected the local minima forming the smoothest reconstructions on each optical window in the sense of Section 8.
6.2.1. Results for Ag
6.2.2. Results for CsI
6.2.3. Results for ![](//:0)
O
6.3. Conclusion
For Ag, the average total run time over all 48 wavelengths for method 1 was less than for method 2, for
O
and for CsI
, i.e. method 1 is almost two times faster than method 2. The results are of the same quality, since their relative deviations are just small fractions of percentages.
The continuation method approach saves run time significantly with the same quality of the results compared to using the truncation index (Equation6.1(6.1)
(6.1) ) all the time.
7. Nonlinear Tikhonov regularization
So far we have solved the regression problem (Equation4.3(4.3)
(4.3) ) without any regularization, thus the obtained refractive index reconstructions might still be too error-contaminated to be of practical use. A widely used regularization strategy for nonlinear regression problems is Tikhonov regularization, which yields the regularized regression problem
(7.1)
(7.1)
when we apply it on (Equation4.3(4.3)
(4.3) ), cf. [Citation9]. Here
is a regularization parameter and
is an estimate of the sought-after true solution. In many cases, the unregularized problem has a whole set of minimizers, thus the vector
works also as a selection criterion. Now if a reasonable
is found, the regularization parameter
can be determined with the discrepancy principle, i.e.
is computed such that
where is an estimate of the residual of the ‘true’ solution which depends on the noise level
. For this task, monotonicity in the residual of
is established in [Citation9].
The problem of finding a good estimate still remains. In [Citation10], an alternative implementable parameter choice strategy without the need of an
is derived. Applied on our problem, it gives
This method has the drawback that the matrix needs to be inverted, which may lead to instabilities.
Nevertheless the quality of the regularized solutions still depends strongly on the start values for solving (Equation7.1(7.1)
(7.1) ). We know about our sought-after refractive indices that they form smooth curves on each of the five optical windows. The complex refractive index curves of most materials can be described using the so-called Lorentz-oscillator-model, cf. [Citation11]. Here points with bigger curvatures only occur at so-called resonance frequencies corresponding to some isolated resonance wavelengths. Motivated by these facts, we derive in the following a method to find reasonable start values for Phillips–Twomey regularization out of the results of Algorithm 1, which will be outlined in Section 9.
8. Finding the smoothest coupled solutions
We have the problem of identifying the best approximation to the sought-after true particle material refractive index out of a set of multiple solutions obtained with Algorithm 1. This problem can be solved by coupling the solutions, which means that we combine solutions from neighbouring wavelengths l in each of the five optical windows in order to obtain a unique solution for every optical window. We know about the complex refractive index curves to be retrieved that they are smooth; hence, we expect their sum of the squared second finite differences both in the real and imaginary parts to be small.
Let , ...,
denote the wavelengths of any of our five wavelength ranges. Let
, ...,
be the number of solutions found for all the wavelengths. We denote with
the jth solution found for wavelength
for
and
. Now we wish to find the smoothest combined solution from all possible combinations
, ...,
for
, hence we have a total number of
combinations. Here we measure smoothness of a combination
, ...,
with the sum
of its second finite differences both in the real parts and its imaginary parts
, which means that we regard a combination the smoother the smaller its sum S is.
We encounter the problem that the total number of possible combinations might get too big to iterate through all combinations in the search for the smoothest one in acceptable time. Therefore we propose a greedy algorithm, which uses each second finite difference as start point to find a smooth combination.
The main loop spanning over the lines 5–52 iterates through all positions of a middle point for a second finite difference both in the real and imaginary part. At each position z, the inner loops beginning in lines 6–8 iterate through all possible second finite differences which can be formed out of the vectors
,
and
for
,
and
, i.e. they loop through all of their middle points and left and right neighbours at position z. In lines 9–12, the variable
is initialized with the sum of the squared second finite differences in the real and imaginary parts of the current vectors
,
and
and the positions
, z and
of the array Comb are filled with the current vectors. For
, the loop in lines 13–28 successively fills the positions
of the array Comb. At each new position k, the minimal sum of the two squared second finite differences in the real and imaginary part
is determined in lines 19–25, where the middle and right points are fixed and taken as the leftmost two vectors from the array Comb and the right point runs through all
for
. After the vector
giving out of all of the
’s giving the minimal sum
is found, the
is added to
and
is stored in kth entry Comb(k). In a similar way, the loop in lines 29–44 succesively fills the positions
for
. This time the left and middle points are fixed and taken as the rightmost two points of the array Comb, whereas the left point iterates through all
for
. Again the vector
is that one of the
’s giving the minimal sum
and it is stored in Comb(k). As well the sum
is added to
.
In the above procedure, every triple of neighbouring vectors from the results of Algorithm 1 is considered to possibly lie on the sought-after smoothest combination with the smallest sum of all squared second differences. The three vectors are used as start points to find a smooth combination with a greedy strategy, where only a vector is added to the current combination, if it gives the smallest sum at the left or right end of the growing set with vectors already added, until the first and last position are reached.
Finally, from all of the combinations constructed this way the smoothest one with the smallest sum out of all
’s is selected in lines 45–48 to be the final output SmoothestCombination.
Define . Then the total number of operations needed for Algorithm 2 can be estimated by
which is considerably less than the
operations needed by the naive method of iterating through all possible combinations.
9. Further regularization of coupled solutions
Not only for determining the smoothest refractive index curve reconstructions formed from the results of Algorithm 1 the coupled view on the solutions is beneficial – it also leads to further improvement of the results by Twomey-regularization. Let us investigate the coupled approach in a probability theoretical setting. Here we reuse the notations introduced in Section 4, i.e. we let , ...,
denote a set of solution for any of the five optical windows. Then the joint posterior probability density of
, ...,
is given by
(9.1)
(9.1) where
is the data vector for the jth wavelength
having N entries with N being the size of the radius grid. Moreover,
is the scaled covariance matrix for
and
is the applied model depending on
. Note that we initially have differing truncation indices
, ...,
. Since the coefficient functions of the truncated model function
are decaying fast for each
, it is convenient to change to the same truncation index
for all wavelengths
, ...,
. The errors introduced by doing so are negliglible. It is easy to show that maximizing the joint density (Equation9.1
(9.1)
(9.1) ) is equivalent to maximize all single densities
independently, i.e. a joint MAP-estimator
consists of the single MAP-estimators
for . This means the the results of Algorithm 1 can be used to construct MAP-estimators for the joint posterior probabilty density.
This behaviour changes when we replace the joint prior probabilty density
with
where
where is a regularization parameter and
is a parameter specifying the amount Tikhonov regularization.
In the new prior distribution, we use a combination of Tikhonov and Phillips–Twomey regularization both in the real and imaginary parts. Here we apply a small amount of Tikhonov-regularization by setting , such that the resulting regularization operator gets regular. This means that the regularized regression problem (Equation9.2
(9.2)
(9.2) ) can be transformed into standard Tikhonov form and that the monotonicity results from [Citation9] are still valid. Each second finite difference is clearly a function of three neighbouring points; therefore, a decoupled computation of the joint MAP-estimator
(9.2)
(9.2)
with
for each wavelength seperately is not possible anymore after changing to the new prior density. However the result vectors , ...,
from Algorithm 2 form a good start vector to solve the nonlinear regression problem (Equation9.2
(9.2)
(9.2) ).
We selected the regularization parameter using the discrepancy principle, i.e. we compute
such that the regularized solution
fulfils a relation of the form
where is a proposed residual value depending on the noise level
. In [Citation2], several different residual values are proposed for a fixed model discretization and a set of regularization parameters is obtained from those using the discrepancy principle. The pairings of model discretizations and regularization parameters obtained this way are compared by their Bayesian posterior probabilities. For these probabilities, a set of integrals of the different model posterior densities is needed to be computed, which can be done with Monte Carlo integration methods. Due to the highly nonlinear behaviour of our model
such integration methods are not available here. Therefore, we simplified the posterior exploration in such a way that only one residual value is proposed.
Since each observed probability density for
is Gaussian, the joint observed density
is Gaussian as well. We have
, thus the sum of residuals
running through all wavelengths in the optical window is
-distributed. This yields
Now a widely proposed residual value for the discrepancy principle is , where
is the so-called Morozov safety factor. This choice is prone to under- or overregularization since the residual value corresponding to the ‘true’ solution might be much smaller or bigger than
. Therefore, we proposed a residual value which depends more dynamically on the observed behoviour of the residual.
Let , ...,
denote the unregularized solutions, i.e. the results of Algorithm 2. Then their squared residual is given by
. We first proposed
where we selected . This means that the residual of the regularized solution is beginning at
at least increased by the factor
, which avoids underregularization. If it then happens that
with
the proposed residual is most likely too big and overregularization occurs. In this case, we corrected
by setting
with .
10. Numerical results
To see how reliable our proposed reconstruction algortihm is, we performed for each of the scatterer materials Ag, O a numerical study with 100 sweeps through all 48 wavelenghts of the five optical windows with the same settings as in Section 6. We found out that the radii
m,
m and
m contain the most information about the refractive indices. This was found by keeping our 48 wavelengths fixed and comparing the quality of inversion results under varying aerosol particle radii. Bigger radii did not improve the results in our simulations and refractive index reconstructions only using bigger radii even turned out to be too unstable. A more thorough treatment of this problem can be found in [Citation7], where a covariance eigenvalue analysis is used. Although not directly comparable with our study of uncoated particles, the coated radii
m,
m and
m carrying the most information content found in this study are roughly comparable to our radii.
We computed original spectral extinctions
for Ag, O and CsI and added zero-mean Gaussian noise to it in order to obtain the simulated noisy spectral extinctions
The standard deviations were taken to be of the true spectral extinctions. Real experiments using 500 wavelengths were contaminated by Gaussian noise with
of the true spectral extinctions as standard deviations. We expect that switching to 48 wavelengths and thus increasing the time resolution of the measurements will lower the standard deviations to a small percentage. We used a sample size of
to compute each mean
of noisy spectral extinctions.
In the following, the results are presented separately for each of the three materials. The uppermost plot presents the relative errors of the unregularized solutions obtained with Algorithm 2 from the original scatterer refractive indices. The next plot displays the run times of Algorithm 1, which returned all local minima of (Equation4.3(4.3)
(4.3) ). These candidate solutions served as input for Algorithm 2. Then the relative errors of the regularized solutions are presented. Finally, the relative errors of the average of the regularized solutions are shown.
10.1. Results for Ag
10.1.2. Relative errors of the regularized solutions
10.1.3. Relative errors of the average of the regularized solutions
10.2. Results for CsI
10.2.2. Relative errors of the regularized solutions
10.2.3. Relative errors of the average of the regularized solutions
10.3. Results for ![](//:0)
O
10.3.2. Relative errors of the regularized solutions
10.3.3. Relative errors of the average of the regularized solutions
10.4. Conclusion
The severest relative errors can be observed for Ag. For the initial unregularized solutions, they lie between 1 and on average and can go up to ca.
in the extreme cases as one can see in the leftmost subplot for the first optical window. The run times of Algorithm 1 lie between 30 and 50 s in the average case and can rise up to 200 s in the extreme cases. A typical sweep through all 48 wavelengths needed ca. 30 minutes in total and this value was very much the same for all three materials. For Ag, the regularization procedure effectively reduced the relative errors such that they are in the range between 0.5 and
on average and are below
in the extreme cases. Finally, one can see in the last plot that the relative errors of the average of all 100 regularized solutions are all below
.
For CsI, the relative errors of the unregularized solutions are already quite small and lie between 0.03 and on average and rise only up to
in the extreme cases. The run times of Algorithm 1 are typically in the range from 25 to 55 s and are always below 95 s. The regularization of the solutions brought only a small improvement of the results here such that the relative errors did not change much. They are still in the same range from 0.03 to
on average but only reach up to ca.
now. The relative errors of the average of the 100 regularized solutions are between 0.01 and
.
Also for O, the relative errors of the unregularized solutions are comparably small and below
on average and still below
in the extreme cases. Especially, the rightmost subplot for the last optical window shows the biggest relative errors, whereas for all the other optical windows the relative errors are below
on average and below
in the extreme cases. A similar behaviour can be observed for the run times of Algorithm 1. For the first four optical windows, they are between 20 and 45 s on average and below 100 s in the extreme cases, whereas for the last optical window they are between 30 and 170 s on average and can even rise up to 350 s. For
O, the regularization procedure improves the relative errors only slightly for the first four optical windows and even increases them for the last optical window such that they can rise up to ca.
on average and
in the extreme cases. The relative errors of the average of the 100 regularized solutions are virtually zero for the first four optical windows and below
for the last optical window.
11. Higher noise levels
To see how our proposed reconstruction algortihm behaves for higher noise levels, we performed for each of the scatterer materials Ag, O and CsI two numerical studies with 10 sweeps through all 48 wavelenghts of the five optical windows with the same settings as in Section 6. We computed original spectral extinctions
for all wavelengths and added zero-mean Gaussian noise to it in order to obtain the simulated noisy spectral extinctions
for the first study and
for the second. The standard deviations were taken to be and
, respectively, of the true spectral extinctions. We used a sample size of
to compute each means
of noisy spectral extinctions.
For brevity, we only present the relative errors of the average of the 10 regularized solutions.
11.1. Results for Ag
11.1.1. Standard deviation of ![](//:0)
![](//:0)
11.1.2. Standard deviation of ![](//:0)
![](//:0)
11.2. Results for CsI
11.2.1. Standard deviation of ![](//:0)
![](//:0)
11.2.2. Standard deviation of ![](//:0)
![](//:0)
11.3. Results for ![](//:0)
O
11.3.1. Standard deviation of ![](//:0)
![](//:0)
11.3.2. Standard deviation of ![](//:0)
![](//:0)
11.4. Conclusion
Whereas the relative errors for CsI and O are still below
, they can rise up to ca.
for Ag. Therefore, the reconstructed refractive indices for Ag under this noise level are most likely not of practical use. This shows that the FASP measurements of monodisperse aerosols must be sufficiently accurate in order to retrieve the scatterer refractive indices from them.
12. Numerical study
We performed four numerical studies for two-component aerosols with log-normal, RRSB and Hedrih model size distributions as outlined in [Citation2]. The aerosol particles were assumed to be homogeneously internally mixed, such that only one effective refractive index was retrieved. One component of the simulated aerosols was O with volume fractions of 0, 11, 22, 33, 44, 56, 67, 78, 89 and
. In the first two studies, we simulated mixtures of
O and CsI, where we used the original aerosol component refractive indices for the first study. For the second study, we used the average of the 100 regularized solutions from Section 10. We did the same for the third and fourth studies, but here we simulated mixtures of
O and Ag. In the third study, we utilized the original aerosol component refractive indices and for the fourth the average of the 100 regularized solutions from Section 10.
We applied the same reconstruction methods described in [Citation2] under the same settings, i.e. for each reconstruction we generated 300 artificial noisy measurements for all 48 wavelengths, where the measurement error was simulated as additive zero-mean Gaussian noise. For each wavelength, the standard deviations were taken as of the solutions of the forward problem. In [Citation2], three different regularization methods, namely Tikhonov, minimal first differences and Twomey regularization, were compared and their results turned out to be very similar. Therefore, we only used Tikhonov regularization in the following. The results for the first study were directly adopted from [Citation2].
For every inversion, we computed the -error of the obtained reconstruction relative to the original size distribution and measured the total run time needed for the inversion. The computations were performed on a notebook with a 2.27 GHz CPU and 3.87 GB accessible primary memory.
13. Results for mixtures of ![](//:0)
O and CsI
13.3. Results for mixtures of ![](//:0)
O and Ag
13.4. Conclusion
The resuts of the first and second study only differ by ca. at most and behave very similarly. The same is for the third and fourth studies. These numerical results indicate that 100 FASP measurement sweeps consisting of 300 single measurements with an accuracy as in Section 10 are sufficient to determine aerosol refractive indices in such a quality, that they are suitable for particle size distribution reconstructions for two-component homogeneously internally mixed aerosols using the FASP. The particle radii of the three monodisperse aerosols generated for the refractive indices retrieval need to be
m,
m and
m, respectively.
14. Outlook
It is of interest to investigate if the methods derived in this study can be extended to the case of core-plus-shell aerosols.
Acknowledgements
I thank Graham Alldredge, Ph.D. for proofreading the manuscript, useful recommendations and fruitful discussions on this topic.
Additional information
Funding
Notes
No potential conflict of interest was reported by the author.
References
- Riziq AA, Erlick C, Dinar E, et al. Optical properties of absorbing and non-absorbing aerosols retrieved by cavity ring down (CRD) spectroscopy. Atmos Chem Phys. 2007;7:1523–1536.
- Alldredge G, Kyrion T. Robust inversion methods for aerosol spectroscopy. Inverse Probl. Sci. Eng. 2016.
- Fu Q, Sun W. Mie theory for light scattering by a spherical particle in an absorbing medium. Appl. Opt. 2001;40(9):1354–1361.
- Wiscombe W. Improved Mie scattering algorithms. Appl Opt. 1980;19(9):1505–1509.
- Abromovitz M, Stegun IA. Handbook of mathematical functions with formulas, graphs, and mathematical tables. Washington (DC): Dover; 1972.
- Deuflhard P, Hohmann A. Numerical mathematics I: an algorithmically oriented introduction. Berlin: de Gruyter; 2002.
- Erlick C, Haspel M, Rudich Y. Simultaneous retrieval of the complex refractive indices of the core and shell of coated aerosol particles from extinction measurements using simulated annealing. Appl Opt. 2011;50(22):4393–4402.
- Goody RM, Yung YL. Atmospheric radiation. Theoretical basis. 2nd ed. New York (NY): Oxford University Press; 1989.
- Engl HW, Kunisch K, Neubauer A. Convergence rates for Tikhonov regularisation of non-linear ill-posed problems. Inverse Prob. 1989;5:523–540.
- Scherzer O, Engl HW, Kunisch K. Optimal a posteriori parameter choice for Tikhonov regularization for solving nonlinear Ill-posed problems. SIAM J Numer Anal. 1993;6:1796–1838.
- Quinten M. Optical properties of nanoparticle systems. Weinheim: WILEY-VCH Verlag; 2010.