Book contents
- Frontmatter
- Dedication
- Contents
- Preface
- Introduction
- Part I Inverse Problems
- 1 Bayesian Inverse Problems andWell-Posedness
- 2 The Linear-Gaussian Setting
- 3 Optimization Perspective
- 4 Gaussian Approximation
- 5 Monte Carlo Sampling and Importance Sampling
- 6 Markov Chain Monte Carlo
- Exercises for Part I
- Part II Data Assimilation
- 7 Filtering and Smoothing Problems and Well-Posedness
- 8 The Kalman Filter and Smoother
- 9 Optimization for Filtering and Smoothing: 3DVAR and 4DVAR
- 10 The Extended and Ensemble Kalman Filters
- 11 Particle Filter
- 12 Optimal Particle Filter
- Exercises for Part II
- Part III Kalman Inversion
- 13 Blending Inverse Problems and Data Assimilation
- References
- Index
5 - Monte Carlo Sampling and Importance Sampling
Published online by Cambridge University Press: 27 July 2023
- Frontmatter
- Dedication
- Contents
- Preface
- Introduction
- Part I Inverse Problems
- 1 Bayesian Inverse Problems andWell-Posedness
- 2 The Linear-Gaussian Setting
- 3 Optimization Perspective
- 4 Gaussian Approximation
- 5 Monte Carlo Sampling and Importance Sampling
- 6 Markov Chain Monte Carlo
- Exercises for Part I
- Part II Data Assimilation
- 7 Filtering and Smoothing Problems and Well-Posedness
- 8 The Kalman Filter and Smoother
- 9 Optimization for Filtering and Smoothing: 3DVAR and 4DVAR
- 10 The Extended and Ensemble Kalman Filters
- 11 Particle Filter
- 12 Optimal Particle Filter
- Exercises for Part II
- Part III Kalman Inversion
- 13 Blending Inverse Problems and Data Assimilation
- References
- Index
Summary
In this chapter we introduce Monte Carlo sampling and importance sampling. These are two general techniques for estimating expectations with respect to a given pdf π. Monte Carlo generates independent samples from π and combines them with equal weights, whilst importance sampling uses independent samples, weighted appropriately, from a different distribution. In quantifying the error in Monte Carlo and importance sampling, we will use a distance on random probability measures that reduces to total variation in the case of deterministic probability measures; and we will introduce the χ2 divergence.
- Type
- Chapter
- Information
- Inverse Problems and Data Assimilation , pp. 59 - 72Publisher: Cambridge University PressPrint publication year: 2023
- 1
- Cited by