Idp Gmes ch05
Idp Gmes ch05
5
Intelligent Gamma-Ray Data Processing for
Environmental Monitoring
storage of radioactive wastes, accidental loss of industrial and medical radiation sources, etc. as well
as nuclear weapons incidents or tests.
Gamma‐ray surveys provide a baseline against which man‐made contamination can be estimated.
This makes the radiological assessment of the environment possible, showing general regional trends
in radionuclide distribution, etc. They may be used to estimate and assess the terrestrial radiation
dose to the human population and to identify areas of potential natural radiation hazard, to
determine the effectiveness of cleanup efforts after a spill or an industrial accident, to trace and to
clean up old contamination, and to verify that new contamination is not occurring. For example,
surveys are regularly carried out in the radiation risk zones around facilities where radioactive
materials are used or produced, such as nuclear power plants and other facilities dealing with
nuclear fuel cycle, radioactive waste and material repositories and their transportation routes,
science, industrial, and medicine sites, etc. Similar surveys are used to assess the contamination in
areas of former mining. Besides territories, of interest is gamma‐ray monitoring of objects that use
radioactive materials or fuel, contaminated buildings, containerized waste, industrial processes,
laboratories, etc., as well as moving objects (vehicles, movable facilities, cargo, people, etc.).
Gamma‐ray monitoring is also closely related to security, which in this context is mainly associated
with the dangers of nuclear proliferation and, more recently, radiological and nuclear terrorism.
Nuclear terrorism denoting the use, or threat of the use, of nuclear or radiological weapons ("dirty
bomb") in acts of terrorism is nowadays becoming an immediate challenge for the entire world
[Ruff, 2006], [Markman et al, 2003]. The challenges of coping with the consequences of possible acts
of nuclear terrorism or of the detection of radiological or nuclear materials in a planted device are
closely related to the issues of environmental gamma‐ray monitoring and spectroscopy mentioned
above.
This also concerns the challenges of pre‐empting nuclear and radiological terrorism that requires
detection, interdiction, and identification of smuggled nuclear materials, as well as preventing or
intercepting the illicit trafficking, theft and loss of nuclear and other radioactive materials and
weapons. Monitoring is required at the borders and ports of entry, including land, rail, marine, or air
border crossings, as well as the facilities where radioactive materials and weapons are produced or
stored. To detect illicit transport of radioactive material in vehicles, people, luggage, mail, cargo,
containers, etc., radiation portal monitors are commonly used.
All those mentioned pressing challenges and applications require ever‐enhancing accuracy, precision,
stability, and reliability of gamma‐ray monitoring. In its turn, this requires novel gamma‐ray
spectroscopic techniques to employ state‐of‐the‐art approaches and methods of intelligent data
processing and machine learning. In this chapter, we discuss gamma‐ray spectrometry techniques for
both close‐range and remote monitoring. First, we consider gamma‐ray spectroscopy with
multichannel spectrometer both for reproducible and unknown geometry of measurements. Then,
techniques for airborne gamma‐ray spectrometry are discussed. Results of experimental comparison
are provided to illustrate the performance of the employed machine learning techniques.
energies and are used to monitor the gross level of the gamma radiation field and to detect the
presence of anomalous sources. Gamma‐ray spectrometers, on the other hand, measure both the
intensity and energy of radiation, thus enabling the source of the radiation to be diagnosed [IAEA,
2003].
In order to interpret spectrometer measurements in real‐world conditions, a number of intelligent
data processing techniques have been developed or adapted. In this section, we review available
techniques, as well as machine learning approaches such as model selection, sparse approximation,
and blind source separation that we develop and/or apply to gamma‐ray spectroscopy. Performance
of those techniques is illustrated using artificial and real gamma‐spectrum data.
Gamma‐ray spectroscopy is the quantitative study of the energy spectra of gamma‐ray sources. The
result of the gamma‐ray spectrum measurement by a multi‐channel spectrometer (Figure 78) may be
considered as L numbers zi (i=1,...,L) each representing the count f(zi) of gamma‐rays of particular
energy zi. This output signal of the spectrometer can be decomposed as
The task is complicated by fluctuations in the natural radiation background, low levels of radionuclide
activity, a complex spectral composition of radiation.
350
Count Cs-137 E=662
Mn-58 E=834
300
Co-60 E=1173
250 Co-60 E=1332
Background
200
Measured
150 Model
100
50
Channel
0
0 50 100 150
Figure 79. Detector response functions (200 functions Figure 80. Spectrum decomposition using detector
of 256 channels each response functions
Measurements in a non‐fixed geometry (unknown and complex) have the following peculiarities:
position of the detector relative to the radiation source and the geometry of the source are
unknown; the radiation source is shaded by a substance that absorbs radiation, geometry and nature
of this screen are unknown.
So, the detector response functions for a non‐fixed geometry of measurements cannot be generated
in advance, and the decomposition of Eq(1) cannot be immediately applied to this case. To cope with
this problem, we proposed [Zabulonov et al, 2004a] to split each detector response function into
three separate functions ψА(z), ψС(z), ψR(z) corresponding to three characteristic spectrum regions:
the total absorption peak, the Compton part and the reverse flight peak. So, each detector response
function is represented as
w = + f, (3)
where is the pseudoinverse matrix.
+
However, due to the noise, this method assigns non‐zero weights to nuclides actually not present in
the spectrum. To increase noise resistance, maximum noise fraction (MNF) [Dickson and Taylor,
1998; 2000] and noise‐adjusted singular value decomposition (NASVD) [Minty and McFadden, 1998],
[Minty and Hovgaard, 2002], [Mauring and Smethurts, 2005] are sometimes used. Regretfully, those
techniques require information about noise.
For a non‐fixed geometry of measurements, performance of the traditional methods is even poorer
than for a fixed geometry. If one uses Eq.(1), detector response functions should be known for all
possible geometries and their proper set should be chosen in‐situ, that cannot be implemented in
practice. If one considers Eq.(2), the drawbacks of the traditional methods described above become
more severe since the number of weights (activities) to be estimated grows.
Another approach to the non‐fixed geometry spectroscopy consists in the following. Some
hypotheses about the composition and activity of nuclides in the measured spectrum are set up.
Then the Monte Carlo simulation of photon transport is done for those hypotheses, taking into
account the particular geometry of measurements. If the measured spectrum is exactly
reconstructed by the simulation, the desired composition and activity are found [Gal et al, 2001],
[Toubon et al, 2006] This approach has a large computational complexity and so cannot provide real‐
time results, since a lot of hypotheses should be verified in case of poor a‐priori information.
Intelligent Data Processing in Global Monitoring for Environment and Security 141
Machine Learning
Detector response
Model Selection
functions for
approach
radionuclides
3
Multi-channel Model Selection Activities
spectrum of present radionuclides of present
measurements using Sparse Approximation radionuclides
To improve the accuracy and stability of radionuclide activity estimation, we develop an approach to
full spectrum processing using machine learning methods of model selection, sparse approximation,
and blind source separation (Figure 82).
Model selection is a problem arising in the areas of machine learning and data mining. Given data
usually consisting of input‐output pairs, a model is built to relate the input and the output. A number
of approaches have been proposed to build the model, including linear models, neural networks,
classification and regression trees, kernel methods, etc. The task of model selection is to choose a
close‐to‐optimal model in terms of the particular model usage and application.
A number of model selection methods use various model selection criteria [Akaike, 1974],
[Mallows, 1973], [Hansen and Yu, 2001]. Model selection criteria are formulated in such a way that
they balance the model complexity vs. the approximation error, and so automatically reduce the
model complexity with increasing noise levels. At the optimal models, criteria reach minimum. Model
selection criteria have been developed based on various assumptions, i.e. using predictive training
error [Mallows, 1973]; generalization error [Sugiyama and Ogawa, 2001], [Niyogi and Girosi, 1996]
and related criteria, such as based on information statistics [Akaike, 1974], description length
[Rissanen, 1978; 2002], etc.
Let the data set DL be represented by L pairs DL={(xi,yi)}i=1,...,L, XLN, yi=y0i + i, is Gaussian additive
noise. The predictive training error is adopted as the error measure [Sugiyama and Ogawa, 2001]
between estimated and true values at sample points contained in the training set:
LyMDL = 0.5(yTy – FSS)/σ2 + 0.5 s1 [1+ log(FSS/ (σ2s))]+ 0.5log(L) for FSS>sσ2,
(10)
L(y) = 0.5(yTy) σ2 otherwise,
gMDL = 0.5L log(RSS/(L–s1)) + 0.5s log(F) + log(L) for R2 ≥ s/L,
(11)
gMDL = 0.5L log (yTy/L) + 0.5 log(L) otherwise,
where FSS = yTXS(XSТXS)–1 XSТy, F= (L–s)(yTy–RSS)/ (s RSS), R is the usual multiple correlation
coefficient.
For the spectroscopy task, the model selection approach is assumed to avoid the situation when the
model includes elements actually not present. In order to use the approach, we must generate the
models and choose the one with the minimal model selection criterion value. Since exhaustive search
is impossible for practical settings, we use greedy techniques based on the ideas of sparse
approximation [DeVore, 1998], [Donoho et al, 2004].
The field of sparse approximation is concerned with the problem of representing a target vector
using a short linear combination of vectors drawn from a large, fixed collection called a dictionary
This problem arises in applications as diverse as machine learning, signal and image processing,
Intelligent Data Processing in Global Monitoring for Environment and Security 143
imaging sciences, communications, numerical analysis and statistics. It is known that sparse
approximation is computationally hard in the general case. Nevertheless, it has recently been
established that there are tractable algorithms for solving these problems in a variety of interesting
cases. It was also shown that, under certain conditions, sparse approximation and Support Vector
Machine technique proposed by V. Vapnik give the same solution [Vapnik and Chervonenkis, 1970].
Sparse approximation methods are used to solve w = y0 for the complete basis (i.e. the basis that
exactly represents y0) with the maximum sparseness in terms of the l0 norm (w*=min||w||0). They
include step‐wise regression [Seber, 1977], k‐term approximation [Temlyakov, 2003], matching
pursuit [Mallat and Zhang, 1993], [Tropp and Gilbert, 2007].
Matching pursuit at the (k+1)‐th step calculates the approximation of y, fk+1=fk+wk+1k+1 choosing k+1
(k+1 without k) and wk+1 that minimize the residual sum squared ||Rk–w||2:
wk=k+ y. (14)
Calculate a new residual:
The relative error was estimated for the radionuclide with the lowest energy 137Cs, since its spectrum
is "hidden" by the other radionuclides. Model selection criteria of Mallows, Akaike, Bin Yu were used
with MPCR. Slightly better results were obtained for the Mallows criterion Cp. The experimental
results are given in Table 3.
Table 3. Relative error of 137Cs activity in the multi‐nuclide sample obtained by ROI, STR, MPCR
When processing the spectrum of known composition (137Cs, 40K), the ROI method provides the best
relative error, as it was set up precisely for those nuclides and is influenced by noise only inside its
regions of interest. MPCR provides a smaller error compared to STR. When processing the spectrum
with additional radionuclides (Th or Th, Ra) unknown to the ROI and STR methods, MPCR provides
the smallest relative error measuring of the 137Cs activity.
For the first experiment in a non‐fixed
geometry of measurements, the samples
Count Cs-137 E=662
including 60Со(Cobalt) (2 spectrum lines),
137
100 Cs-134 E=795
Cs, 134Cs, 58Mn (Manganese) have been
Mn-58 E=834
simulated by their 3‐part detector response 80
functions Eq.(2) (Figure 83). The weight of Co-60 E=1173
the reverse flight peak was 0 and the weight 60 Co-60 E=1332
true activity vector and the estimated activity vector at different noise levels. The results in Figure 84
show that MPCR for all model selection criteria and all noise levels performs better than FULL, but
worse than TRUE. The error for MPCR with MDL is somewhat less than that for CP and AIC.
The next experiment was conducted using the
really measured spectra of a point gamma
700
Error Cp
radiation source on the background of a MDL
distributed source with a higher energy. ROI, 600
AIC
STR, and MPCR with the MDL model selection 500 FULL
criterion were compared. In the experiment, the TRUE
400
point source was 137Cs and the distributed source
was 40K; the 40K activity was 51% of the 137Cs 300
activity. ROI and STR were set up for 137Cs and 200
40
K. The relative error of the 137Cs activity was
92% for ROI, 85% for STR, and 12% for MPCR. 100
Noise Level
0
Model optimality tests 0 10 20 30 40 50 60 70 80 90 100
In order to determine why different model Figure 84. Radionuclide activity error vs noise level
selection criteria vary in accuracy, we compared for various models and model selection criteria
[Revunova, 2008], [Zabulonov et al, 2004a],
[Zabulonov et al, 2009a] the dependence of the model dimensionality and the model error on the
noise level. The comparison showed that the best accuracy is provided by the criterion that keeps the
true dimension of the model at the higher noise levels than the other model selection criteria.
In [Gribonval et al, 2006] a test was proposed for checking if the particular set of basic functions in
the model is true. It is based on the notion of l0‐optimal solution that provides both the minimum
approximation error and the maximum model sparsity. Experiments showed [Revunova, 2008] that
with the increasing noise levels the error using such a test is smaller than that of the traditional
model selection criteria [Revunova, 2005a], [Revunova and Rachkovskij, 2005], [Zabulonov et al,
2005].
The drawback of this model optimality test is that it can only be used for some bases with valid "basis
connectivity" condition [Gribonval et al, 2006]. So, research on how to solve this problem for certain
types of detector response basis functions could lead to further improvement in the accuracy of
radionuclide activity estimation under noisy conditions.
Blind source separation (BSS) is the separation of a set of signals from a set of mixed signals, without
the aid of information (or with very little information) about the source signals or the mixing process.
In machine learning, BSS is considered as one of the most powerful unsupervised learning
techniques. Unsupervised learning is distinguished from supervised learning and reinforcement
learning in that the learner is given only unlabeled examples. Theoretical foundations of the BSS
methods are well developed [Chichocki and Amari, 2002], [Comon, 1994], [Hyvarinen, 1999].
The most straightforward case is linear BSS with the number of observed mixtures equal to the
number of unknown sources. Here the data DL are represented by the set of L mixture vectors yiN:
DL={yi}i=1,..,L. The mixture vector y depends on the source vector x as a linear model у=Aх, where the
146 Intelligent Data Analysis in Global Monitoring for Environment and Security
full rank mixture matrix ANN is unknown. The task is to obtain L unobserved vectors x (matrix
XLN) using L known vectors y (matrix YLN) by estimating the un‐mixture matrix
B=(b1,…,bN)TNN and then calculating X=YB.
Well‐known approaches to solution of the BSS problem use a priori information about the sources.
The sources are supposed to be statistically independent and having nongaussian distribution
(independent component analysis ICA [Amari and Cichocki, 1998], [Amari et al, 2002]), or
uncorrelated and having the Gaussian distribution (principal component analysis PCA [Chichocki and
Amari, 2002]), or sparse (sparse component analysis SCA [Li et al, 2004]), etc. Based on those
assumptions, a cost function is constructed (e.g., mutual information of sources) and then minimized
by the optimization algorithms, providing the solution.
Usually, BSS‐related methods such as MNF [Dickson and Taylor, 1998; 2000] are used in spectroscopy
for noise suppression. We propose using BSS in spectrometry to reconstruct detector response
function as follows [Zabulonov et al, 2009b]. A system of several detectors located at various
distances from the radiation sources simultaneously measures the gamma radiation spectra. Those
multiple measured spectra are processed using the BSS method and at the output we get the
separated detector response functions of individual radionuclides as reconstructed mixture
components. This BSS approach to spectroscopy does not require measurement of the background
radiation, knowledge of detector response functions, and the number of reconstructed hidden
sources only depends on the number of detectors.
We carried out experiments [Zabulonov et al, 2009b] investigating BSS performance in spectrometric
tasks. In one of the experiments, the radiation sources Cs137 and K40 were measured simultaneously
by three detectors located at various distances from the sources.
The measured spectra (Figure 85) were input to the Independent Component Analysis algorithm
FastICA [Hyvarinen, 1999] belonging to the BSS family. The separated hidden sources corresponding
to 137Cs, 40K and background radiation are presented in Figure 86.
mix 1 7 Cs-137
Count
60 mix 2 K-40
6
mix 3 Background
50 5
40 4
3
30
2
20
1
10 0
Channel Channel
0 -1
0 50 100 150 0 50 100 150
Figure 85. The results of gamma‐ray spectra Figure 86. Independent components separated from
measurements by 3 detectors. 137Cs, 40К, and are the spectrum mixtures: 137Cs, К40, and background
mixed in the spectra
The BSS methods are sensitive to noise and require large data sets, so that their usage in non‐
stationary applications is rather problematic [Bach and Jordan, 2002]. To improve BSS performance
Intelligent Data Processing in Global Monitoring for Environment and Security 147
in this situation, we have developed a cost function based on the algorithmic complexity approach,
and algorithmic mutual information in particular [Grunvald and Vitanyi, 2003]:
random projections, pseudoinverse, and show that its accuracy is comparable to the Tikhonov
regularization, but is less expensive computationally.
not only on the ground radioelement concentrations, but also on the equipment used, and on the
nominal height of the survey. This is undesirable, as measuring units should have a direct significance
and be independent of the instrumentation and survey parameters. Count rates should therefore be
converted to ground concentrations of the radioelements.
The task is to restore the radionuclide surface density function g(х,у) by the detector readings
F(хd,уd). Due to the large angular aperture of the collimator, the measurement results (detector
readings, count rates) are smoothed and do not reflect in detail the surface density pattern (Figure
88).
Figure 88. The surface density function (left) and its measurements (right)
Let us consider intelligent information processing techniques for reconstruction of the surface
density. The detector count rates are related to the unknown surface density by a Fredholm integral
equation of the first kind. For the case of stationary detector placed above the point (xd,yd) at the
height hd this equation is
Figure 89. Deconvolution and regularization machine learning approaches to airborne gamma‐ray surveying
The deconvolution approach has the following drawbacks. Derivation of the Wiener filter requires
knowledge of spatial statistics of both the actual distribution density g(x,y) and the measurement
noise (x,y). Since both are unknown, some assumptions should be made, that may be not always
Intelligent Data Processing in Global Monitoring for Environment and Security 151
valid. For example, determining the noise levels after multichannel processing such as NASVD, and
particularly after application of the adaptive filter, is difficult. So, in [Billings et al, 2003], after getting
a ballpark estimate of the noise, manual adjustments are used to visually tune the deconvolution.
Also, the blurring in the data is assumed to be the same everywhere. This means that no
accommodation can be made for changes in the detector height or for 3D terrain effects.
The noise is assumed additive and the same everywhere. Due to the Poisson nature of radioactive
decay and gamma‐ray detection, this is clearly not the case. To mitigate this effect, sometimes an
adaptive 2D Lee filter [Ristau and Moon, 2001] is used to remove random noise from an image while
maintaining image edges.
In situations where the terrain clearance (variations in survey altitude and topography) and noise
vary significantly over the survey area, a partial solution would be to break the survey up into smaller
segments and apply deconvolution separately to each sub‐area.
An alternative approach [Zabulonov et al, 2006] is to use a full space domain formulation. For a
particular example, let us consider a detector whose field of view is defined by a vertically oriented
collimator of the horn type with a rectangular cross section and angular aperture (x,y),
e.g. 0.5x = 0.5y 60. Therefore, the detector's field of view looks at a square piece of the surface
topography, called "window" hereafter. The window size is determined by the angular aperture of
the collimator and the height of detector. The window movement during the exposure time makes a
rectangular exposure strip.
This setup is described by Eq.(18), where integration is over the area of the browsing window on the
ground surface Sd, and
5.3.3 Regularized solution of the ill‐posed inverse problem for a full spacedomain
formulation
Aх=b, (25)
where the matrix AN×N and the vector bN are obtained from discretization of the kernel K and
f(xd), respectively; x represents the unknown g(х,у=const) along the flight line to be reconstructed.
The vector bN is distorted by the additive noise N: b=b0+.
In case when the singular values i of A decay gradually to zero, the ratio between the largest and
the smallest nonzero singular values is large, the problem is known as discrete ill‐posed problem
[Hansen, 1998]. Actually, these properties are present in the spectrometry task considered here.
Approximate solutions of discrete ill‐posed problems as the least squares problem
152 Intelligent Data Analysis in Global Monitoring for Environment and Security
The drawbacks inherent in the methods of solving discrete ill‐posed problems based on Tikhonov
regularization include their high computational complexity and the difficulty of selecting the proper
regularization parameter (penalty weight) which influences the solution stability. At the wrong values
of the regularization parameter, the error of solution may be substantial. Therefore, alternative
approaches are required for solving discrete ill‐posed problems that would have the accuracy
comparable to Tikhonov regularization at lower computational costs.
We develop such an approach using the ideas of random projections [Johnson and Lindenstrauss,
1984], [Papadimitriou et al, 2000], [Achlioptas, 2003] and of our previous work on distributed
representations (e.g., [Misuno et al, 2005]) inspired by the idea of information representation in
neural networks of the brain. Random projections have recently appeared as a tool for
dimensionality reduction and have been used to produce a number of interesting results, both
theoretical and applied, including those in context of inductive supervised learning using machine
learning methods [Halko et al, 2009], [Revunova and Rachkovskij, 2009s].
The technique plays a key role in several breakthrough developments in the field of algorithms. In
other cases, it provides elegant alternative proofs. Recently, researchers working in the area of
numeric linear algebra applied similar ideas to get fast randomized algorithms for the least squares
problem, matrix factorization, principal component analysis, etc. [Sarlos, 2006], [Drineas et al, 2007],
[Rokhlin and Tygert, 2008], [Tygert, 2009], [Halko et al, 2009]. It is therefore of interest to study
those techniques and apply them to discrete ill‐posed inverse problems.
Let us use the randomized algorithms not only to accelerate, but also to stabilize the solution x' of
the ill‐posed problem, as follows [Revunova and Rachkovskij, 2009s]. Multiply both sides of Eq.(25)
by the matrix K×N, K≤N, whose elements are realizations of a normal random variable with zero
mean and unit variance. The number of columns N of matrix is determined by the dimension of
the matrix A, the number of rows K is a priori unknown since the numerical rank of A is ill‐
determined and the required numerical rank of approximation is unknown. We obtain
A x = b, where A K×N, bK. (29)
Then the least‐squares problem is
хPr = argminх ||A x – b||2. (30)
Signal reconstruction based on pseudo‐inverse using random projection is obtained as
хpinPr = (A)+ b. (31)
Signal reconstruction based on the pseudo‐inverse using the projection matrix Q obtained by the QR
factorization of A is done as
хpinQ= (QTA)+ QTb. (32)
The pseudo‐inverse P+ of a matrix P is actually computed based on SVD as:
P+ = V diag (φi / σi) UT, iff σi>tresh φi=1, otherwise φi =0.
(33)
tresh = max(K,N) eps(max(σi)),
where U, V, S are obtained by the SVD of P = USVT; σi = diag S are singular values, the elements of a
diagonal matrix S; floating‐point relative accuracy eps(z) is the positive distance from abs(z) to the
next larger in magnitude floating point number of the same precision as z.
154 Intelligent Data Analysis in Global Monitoring for Environment and Security
In order to compare the quality of reconstruction of the surface density of radioactive sources, we
conducted an experimental study of techniques for solving discrete ill‐posed problems using the
simulated data of radioactivity monitoring at h=100m.
The matrix (i.e., A in Eq.(25)) obtained by discretization of the kernel Eq.(24) has dimensionality of
200×200 (Figure 90), large condition number (max/min>>1), and singular values i gradually decaying
to zero. The right‐hand side y (i.e., b in Eq.(25)) is distorted by an additive noise with the Gaussian
distribution and various amplitudes. Figure 91 shows the measurement y produced by the doublet
signal x to be restored.
6E+61 5
Decays Count
5E+61 4
x
4E+61 x*
y 3
3E+61
2
2E+61
1
1E+61
Coordinate
0 0
-1E+61 -1
Figure 90. The kernel matrix for gamma‐ray Figure 91. The modeled 1D surface density х, the
surveying measurement y, and the restored density x* for
gamma‐ray surveying
First, we have compared performance of three kinds of techniques: the pseudo‐inverse solution
Eq.(26), the Tikhonov regularization Eq.(28) with the selection of the regularization parameter
using the L‐curve, the generalized cross‐validation, and the discrepancy principle [Hansen, 1998],
[Wahba, 1990], [Morozov, 1984], and the pseudo‐inverse with projection using the random matrix
Eq.(31) and with the ortogonalized matrix Q Eq.(32). The random projection matrix K×N, N=200,
K≤N is the Gaussian random matrix with the entries of zero mean and unit variance. Three noise
levels are used {10–2,10–6,10–10}. The results in terms of the signal reconstruction error vector norm
are given in Table 4.
Without projection, the standard pseudo‐inverse provides an acceptable error ePin only at the lowest
noise level and cannot be used at all at the larger noise levels due to very large error. The errors for
the Tikhonov regularization with the selection of by the L‐curve eRegT Lcur, the generalized cross‐
validation eRegT GCV, and the discrepancy principle eRegT Dsc are generally comparable for all three
noise levels. However, we observe an outlier for eRegT GCV at the noise level 10–6, giving the instance
of unstable regularization parameter selection. The lowest signal reconstruction error is provided by
the Tikhonov regularization with the L‐curve.
Intelligent Data Processing in Global Monitoring for Environment and Security 155
Table 4. The signal reconstruction error for the discrete ill‐posed problem of gamma‐surveying
obtained by the regularization methods without and with random projecting. K is the
dimensionality of projection matrix.
With the random projection, the pseudo‐inverse becomes stable and the reconstruction error values
at the minimum (with the proper choice of K) are comparable and small for the projector matrix
(Eq.(31)) ePinR and Q (Eq.(32)) ePinQR.
In order to see how to choose the optimal K for
the projective methods, let us consider the 1.E+08 ePinQR ePinQR 1e-2 6000
gMDL
ePinQR 1e-6
dependence of the signal reconstruction error ePinQR 1e-10
gMDL 1e-2 5000
ePinQR on the dimensionality K of the projector
1.E+06 gMDL 1e-6
matrix Q (Figure 92, left Y‐axis). The error gMDL 1e-10 4000
minimum can be observed for all noise levels.
With increasing noise level, position of the error 1.E+04 3000
5.4 Discussion
In this chapter, we discussed machine‐learning techniques for intelligent processing of gamma‐ray
spectroscopy data in environmental and security‐related monitoring.
In particular, we developed and applied to intelligent processing of multichannel gamma‐ray
spectrometer data the supervised learning techniques of sparse approximation with model selection
that resulted in a substantial increasing of data processing quality. Development of a new model
selection test is suggested for further improving the quality of gamma‐ray spectrometry. For this
task, we also considered blind source separation approach of unsupervised learning, including a
novel objective function based on algorithmic complexity. It is expected to become an important step
in the development of the new generation of intelligent information technologies for multichannel
and multidetector gamma‐ray spectroscopy under a complex and unknown geometry of
measurements.
Intelligent Data Processing in Global Monitoring for Environment and Security 157