Muutke küpsiste eelistusi

Regularization Theory for Ill-posed Problems: Selected Topics [Kõva köide]

  • Formaat: Hardback, 303 pages, kõrgus x laius: 240x170 mm, kaal: 651 g, 24 Tables, black and white; 35 Illustrations, black and white
  • Sari: Inverse and Ill-Posed Problems Series
  • Ilmumisaeg: 17-Jul-2013
  • Kirjastus: De Gruyter
  • ISBN-10: 3110286467
  • ISBN-13: 9783110286465
Teised raamatud teemal:
  • Formaat: Hardback, 303 pages, kõrgus x laius: 240x170 mm, kaal: 651 g, 24 Tables, black and white; 35 Illustrations, black and white
  • Sari: Inverse and Ill-Posed Problems Series
  • Ilmumisaeg: 17-Jul-2013
  • Kirjastus: De Gruyter
  • ISBN-10: 3110286467
  • ISBN-13: 9783110286465
Teised raamatud teemal:
This monograph is a valuable contribution to the highly topical and extremly productive field of regularisation methods for inverse and ill-posed problems. The author is an internationally outstanding and accepted mathematician in this field. In his book he offers a well-balanced mixture of basic and innovative aspects. He demonstrates new, differentiated viewpoints, and important examples for applications. The book demontrates the current developments in the field of regularization theory, such as multiparameter regularization and regularization in learning theory.

The book is written for graduate and PhD students and researchers in mathematics, natural sciences, engeneering, and medicine.
Preface vii
1 An introduction using classical examples
1(46)
1.1 Numerical differentiation. First look at the problem of regularization. The balancing principle
1(11)
1.1.1 Finite-difference formulae
1(2)
1.1.2 Finite-difference formulae for nonexact data. A priori choice of the stepsize
3(3)
1.1.3 A posteriori choice of the stepsize
6(3)
1.1.4 Numerical illustration
9(1)
1.1.5 The balancing principle in a general framework
10(2)
1.2 Stable summation of orthogonal series with noisy coefficients. Deterministic and stochastic noise models. Description of smoothness properties
12(13)
1.2.1 Summation methods
13(1)
1.2.2 Deterministic noise model
14(1)
1.2.3 Stochastic noise model
15(3)
1.2.4 Smoothness associated with a basis
18(1)
1.2.5 Approximation and stability properties of λ-methods
19(2)
1.2.6 Error bounds
21(4)
1.3 The elliptic Cauchy problem and regularization by discretization
25(22)
1.3.1 Natural linearization of the elliptic Cauchy problem
27(9)
1.3.2 Regularization by discretization
36(3)
1.3.3 Application in detecting corrosion
39(8)
2 Basics of single parameter regularization schemes
47(116)
2.1 Simple example for motivation
47(2)
2.2 Essentially ill-posed linear operator-equations. Least-squares solution. General view on regularization
49(16)
2.3 Smoothness in the context of the problem. Benchmark accuracy levels for deterministic and stochastic data noise models
65(15)
2.3.1 The best possible accuracy for the deterministic noise model
68(5)
2.3.2 The best possible accuracy for the Gaussian white noise model
73(7)
2.4 Optimal order and the saturation of regularization methods in Hilbert spaces
80(10)
2.5 Changing the penalty term for variance reduction. Regularization in Hilbert scales
90(11)
2.6 Estimation of linear functionals from indirect noisy observations
101(12)
2.7 Regularization by finite-dimensional approximation
113(11)
2.8 Model selection based on indirect observation in Gaussian white noise
124(17)
2.8.1 Linear models given by least-squares methods
127(4)
2.8.2 Operator monotone functions
131(6)
2.8.3 The problem of model selection (continuation)
137(4)
2.9 A warning example: an operator equation formulation is not always adequate (numerical differentiation revisited)
141(22)
2.9.1 Numerical differentiation in variable Hilbert scales associated with designs
143(4)
2.9.2 Error bounds in L2
147(3)
2.9.3 Adaptation to the unknown bound of the approximation error
150(1)
2.9.4 Numerical differentiation in the space of continuous functions
151(4)
2.9.5 Relation to the Savitzky-Golay method. Numerical examples
155(8)
3 Multiparameter regularization
163(40)
3.1 When do we really need multiparameter regularization?
163(2)
3.2 Multiparameter discrepancy principle
165(12)
3.2.1 Model function based on the multiparameter discrepancy principle
168(2)
3.2.2 A use of the model function to approximate one set of parameters satisfying the discrepancy principle
170(2)
3.2.3 Properties of the model function approximation
172(1)
3.2.4 Discrepancy curve and the convergence analysis
173(1)
3.2.5 Heuristic algorithm for the model function approximation of the multiparameter discrepancy principle
174(1)
3.2.6 Generalization in the case of more than two regularization parameters
175(2)
3.3 Numerical realization and testing
177(12)
3.3.1 Numerical examples and comparison
177(5)
3.3.2 Two-parameter discrepancy curve
182(2)
3.3.3 A numerical check of Proposition 3.1 and use of a discrepancy curve
184(3)
3.3.4 Experiments with three-parameter regularization
187(2)
3.4 Two-parameter regularization with one negative parameter for problems with noisy operators and right-hand side
189(14)
3.4.1 Computational aspects for regularized total least squares
191(1)
3.4.2 Computational aspects for dual regularized total least squares
192(1)
3.4.3 Error bounds in the case B = I
193(2)
3.4.4 Error bounds for B ≠ I
195(2)
3.4.5 Numerical illustrations. Model function approximation in dual regularized total least squares
197(6)
4 Regularization algorithms in learning theory
203(52)
4.1 Supervised learning problem as an operator equation in a reproducing kernel Hilbert space (RKHS)
203(6)
4.1.1 Reproducing kernel Hilbert spaces and related operators
205(2)
4.1.2 A priori assumption on the problem: general source conditions
207(2)
4.2 Kernel independent learning rates
209(9)
4.2.1 Regularization for binary classification: risk bounds and Bayes consistency
217(1)
4.3 Adaptive kernel methods using the balancing principle
218(17)
4.3.1 Adaptive learning when the error measure is known
220(3)
4.3.2 Adaptive learning when the error measure is unknown
223(2)
4.3.3 Proofs of Propositions 4.6 and 4.7
225(6)
4.3.4 Numerical experiments. Quasibalancing principle
231(4)
4.4 Kernel adaptive regularization with application to blood glucose reading
235(14)
4.4.1 Reading the blood glucose level from subcutaneous electric current measurements
242(7)
4.5 Multiparameter regularization in learning theory
249(6)
5 Meta-learning approach to regularization - case study: blood glucose prediction
255(22)
5.1 A brief introduction to meta-learning and blood glucose prediction
255(4)
5.2 A traditional learning theory approach: issues and concerns
259(2)
5.3 Meta-learning approach to choosing a kernel and a regularization parameter
261(8)
5.3.1 Optimization operation
263(4)
5.3.2 Heuristic operation
267(1)
5.3.3 Learning at metalevel
267(2)
5.4 Case-study: blood glucose prediction
269(8)
Bibliography 277(12)
Index 289
Shuai Lu, Fudan University, Shanghai, PR China; Sergei V. Pereverzev, Johann Radon Institute for Computational and Applied Mathematics (RICAM), Austrian Academy of Sciences, Linz, Austria.