SlideShare a Scribd company logo
Slides from Ved Lekic
INVERSE THEORY
1
2
Parameterization
u What quantities
(parameters) should
we consider?
u Consider pairs of
quantities, such as
measurements (y)
made at conditions
(x)
Measurements
(data)
Conditions (independent variable)
Parameterization
u Physical model may
provide insight into
how to parameterize
a problem
u Sometimes we seek
“simplest” model
(fewest parameters)
that can fit data
Measurements
(data)
Conditions (independent variable)
Law of parsimony – a useful heuristic
u Preference for “simpler” models is often justified by invoking “Occam’s Razor”.
u e.g. “It is vain to do with more what you can do with less.”
– some 13th century monk
u Karl Popper argues that simplest
models are preferable because
they can be more easily falsified
u "I remember my friend Johnny
von Neumann used to say, with
four parameters I can fit an
elephant, and with five I can
make him wiggle his trunk.“
-- Enrico Fermi, Nature (2004).
Measurements
(data) Conditions (independent variable)
Under-determined problem
u With 1 data point, we
cannot uniquely
determine slope and y-
intercept
u Complete trade-off
between m1 and m2!
Measurements
(data)
Conditions (independent variable)
Even-determined problem
u When we have 2 data
points, we can uniquely
determine m1 and m2
Measurements
(data)
Conditions (independent variable)
Over-determined problem
u With >2 points, the
line is not guaranteed
to pass through all
points.
u What do we choose?
u What’s the BEST line
… objectively
speaking?
Measurements
(data)
Conditions (independent variable)
The “objective” function
u We introduce an
objective (a.k.a. cost,
misfit) function that is
minimum when our line
provides the “best” fit
u i.e. we minimize εj’s
Measurements
(data)
Conditions (independent variable)
ε4
ε3
ε2
ε1
Choices of Norm for objective functions
u L-1
u Sum of absolute values of residuals
u Tends to ignore outliers
u L-2
u Sum of squared residuals
u Tries to fit outliers
u “Least-squares”
u L-∞
u Maximum residual
u Fit is “never terrible” but usually not good, either
u “Minimax”
φ 𝑚 = $
!"#
$!
𝑑! − 𝑔!(𝑚)
φ 𝑚 = 𝑚𝑎𝑥 𝑑! − 𝑔!(𝑚)
φ 𝑚 = $
!"#
$!
𝑑! − 𝑔!(𝑚)
%
What about data uncertainty?
u If error of one measurement
is expected to be larger than
that of another, then the line
need not pass close to it, nor
should the objective function
count it as much as more
accurate measurements
Measurements
(data)
Conditions (independent variable)
What about data uncertainty?
u If data error is
normally distributed
with variance σj
2 then
the correct objective
function is L-2
u Each squared residual
gets weighted by
variance:
Measurements
(data) Conditions (independent variable)
φ 𝑚 = $
!"#
$!
𝑑! − 𝑔!(𝑚)
%
𝜎!
%
u Matrix notation:
Fitting a line to some points
u Linear problem:
u 𝑦! = 𝑚"𝑥! + 𝑚#
Y
-
Measurements
(data)
X - Conditions (independent variable)
φ 𝑚 = $
!"#
$!
𝑑! − 𝑔!(𝑚)
%
𝜎!
%
𝑥# 1
𝑥% 1
𝑥& 1
𝑥' 1
𝑚!
𝑚"
𝑦!
𝑦"
𝑦#
𝑦$
=
𝐶!
"#/%
𝑑 = 𝐶!
"#/%
𝐺𝑚
⁄
#
(" 0
0 ⁄
#
(#
0 0
0 0
0 0
0 0
⁄
#
($ 0
0 ⁄
#
(%
𝜑 = (𝑑 − 𝐺𝑚)&
𝐶!
"#
(𝑑 − 𝐺𝑚)
⁄
#
(" 0
0 ⁄
#
(#
0 0
0 0
0 0
0 0
⁄
#
($ 0
0 ⁄
#
(%
u Matrix notation:
u Minimize φ with respect to m
u Condition is satisfied when:
u “Least squares” estimate of
model parameters: mest
Fitting a line to some points
𝐶!
"#/%
𝑑 = 𝐶!
"#/%
𝐺𝑚
𝜑 = (𝑑 − 𝐺𝑚)&
𝐶!
"#
(𝑑 − 𝐺𝑚)
'(
')
= 0
𝑚*+, = (𝐺&
𝐶!
"#
𝐺)"#
𝐺&
𝐶!
"#
𝑑
u Linear problem:
u 𝑦! = 𝑚"𝑥! + 𝑚#
Y
-
Measurements
(data)
X - Conditions (independent variable)
φ 𝑚 = $
!"#
$!
𝑑! − 𝑔!(𝑚)
%
𝜎!
%
How confident are we in mest?
u We can evaluate the uncertainty of model parameters using the
posterior model covariance matrix: "
𝐶0
u But first, let’s build some intuition…
Conditions
Measurements
Which scenario – left or
right – will yield a less
uncertain estimate of
slope?
Measurements
Conditions
How confident are we in mest?
u We can evaluate the uncertainty of model parameters using the posterior
model covariance matrix: .
𝐶)
u But first, let’s build some intuition…
Conditions
Measurements
Which scenario – left or right – will yield a less
uncertain estimate of slope?
Measurements
Conditions
BETTER!
à Objective function is more
affected by slope variations
when data uncertainty is small
à more accurate data tends to
yield more reliable estimates of
parameters
How confident are we in mest?
u We can evaluate the uncertainty of model parameters using the posterior
model covariance matrix: .
𝐶)
u But first, let’s build some intuition…
Conditions
Measurements
Which scenario – left or right – will yield a less uncertain estimate of slope?
Measurements
Conditions
BETTER!
à Predictions of widely
spaced (in x) points are more
sensitive to slope variations
à estimates of parameters
to which you are more
sensitive tend to be better!
How confident are we in mest?
u We can evaluate the uncertainty of model parameters using the posterior
model covariance matrix: .
𝐶)
𝑚*+, = (𝐺&
𝐶!
"#
𝐺)"#
𝐺&
𝐶!
"#
𝑑
"
𝐶0= (𝐺&
𝐶!
"#
𝐺)"#
"
𝐶0=
𝜎+-./*
%
𝜎+-./*𝜎01,𝜌
𝜎+-./*𝜎01,𝜌 𝜎01,
%
Variance of slope
estimate
Variance of y-
intercept estimate
Correlation between
estimates of slope and
y-intercept à how
much parameters
trade-off!
Limitation of !
𝐶!
u Posterior model covariance matrix only describes the generalized
Gaussian distribution around the optimal model
𝑚*+, = (𝐺&
𝐶!
"#
𝐺)"#
𝐺&
𝐶!
"#
𝑑
"
𝐶0= (𝐺&
𝐶!
"#
𝐺)"#
"
𝐶0=
𝜎+-./*
%
𝜎+-./*𝜎01,𝜌
𝜎+-./*𝜎01,𝜌 𝜎01,
%
Variance of slope
estimate
Variance of y-
intercept estimate
Correlation between
estimates of slope and
y-intercept à how
much parameters
trade-off!
Schematic representation
data
model
LINEAR PROBLEM
d
= Gm
m
est
after Tarantola, 2005
Low
High
Low
Likelihood
Types of inverse problems
Linear – least squares works and
!
𝐶) is accurate
Linearizable around starting model
!
𝐶) is probably OK
Tarantola,
2005
Non-linear à multi-modal posterior
on m, !
𝐶) is woefully invalid!
• Don’t use least-squares!
SAMPLING!J
Linearizable around most likely
model
• Must iterate!
• !
𝐶) might be OK
Tarantola,
2005
nothing will work! 𝐺𝑜𝑜𝑑 𝑙𝑢𝑐𝑘!
Tarantola,
2005
data
model
d
= Gm
Perfect
Inverse
Problem
m
est
u Unrealistic illustration: Perfect data, perfectly known
G, and all model parameters reconstructed perfectly.
Uncertainty about relationship
between data and model
data
model
d
= Gm
Modeling
Uncertainty
Model
uncertainty!
data
model
Modeling
Error
m
est
d
≈ Gm
m
approx
Model
Error!
Effect of forward modeling error
u Consequences of incorrectly computing g(m) can be
significant:
u Equivalent to errors to data
thereby increasing model
estimate uncertainty
Lekic
and
Romanowicz,
2011
u Can systematically bias
the model estimate
"
𝐶0= (𝐺&
𝐶! + 𝐶&
"#
𝐺)"#
Common complication…
u Often, real-world problems are mixed-determined
u Some parameters are well constrained
u Other parameters are poorly or completely un-constrained
u Sometimes it happens that det(𝐺1
𝐶2
34
𝐺) = 0 and therefore
(𝐺1𝐶2
34𝐺))# – a.k.a. G-g – does NOT exist (i.e. there are an
insufficient number of linearly independent constraints to
determine the unknowns)
u What do we do in that case?

More Related Content

DOCX
Anomaly detection Full Article
PDF
Curvefitting
PPTX
simple and multiple linear Regression. (1).pptx
PPT
Get Multiple Regression Assignment Help
PDF
Prob and statistics models for outlier detection
PDF
Intro to Quant Trading Strategies (Lecture 10 of 10)
PDF
슬로우캠퍼스: scikit-learn & 머신러닝 (강박사)
PDF
Linear Regression
Anomaly detection Full Article
Curvefitting
simple and multiple linear Regression. (1).pptx
Get Multiple Regression Assignment Help
Prob and statistics models for outlier detection
Intro to Quant Trading Strategies (Lecture 10 of 10)
슬로우캠퍼스: scikit-learn & 머신러닝 (강박사)
Linear Regression

Similar to Inversion_Parmetrization_under_det_problem.pdf (20)

PDF
Estimating Causal Effects from Observations
PDF
Linear regression model in econometrics undergraduate
PPTX
2. diagnostics, collinearity, transformation, and missing data
PPTX
Machine learning session4(linear regression)
PPTX
1. linear model, inference, prediction
PDF
Data Science - Part XII - Ridge Regression, LASSO, and Elastic Nets
PPTX
regression analysis presentation slides.
PDF
Big Data Analysis
PPTX
Introduction to Maximum Likelihood Estimator
PDF
Simple linear regression
PPTX
mcp-bandits.pptx
PPTX
Unit III_Ch 17_Probablistic Methods.pptx
PPTX
Bootcamp of new world to taken seriously
PDF
Lect w8 w9_correlation_regression
PDF
MachineLearning_Unit-II.FHDGFHJKpptx.pdf
PPTX
Linear Regression for Data Mining Application
PPTX
Advanced Econometrics L5-6.pptx
DOCX
FSE 200AdkinsPage 1 of 10Simple Linear Regression Corr.docx
PDF
Monte Carlo Simulations (UC Berkeley School of Information; July 11, 2019)
Estimating Causal Effects from Observations
Linear regression model in econometrics undergraduate
2. diagnostics, collinearity, transformation, and missing data
Machine learning session4(linear regression)
1. linear model, inference, prediction
Data Science - Part XII - Ridge Regression, LASSO, and Elastic Nets
regression analysis presentation slides.
Big Data Analysis
Introduction to Maximum Likelihood Estimator
Simple linear regression
mcp-bandits.pptx
Unit III_Ch 17_Probablistic Methods.pptx
Bootcamp of new world to taken seriously
Lect w8 w9_correlation_regression
MachineLearning_Unit-II.FHDGFHJKpptx.pdf
Linear Regression for Data Mining Application
Advanced Econometrics L5-6.pptx
FSE 200AdkinsPage 1 of 10Simple Linear Regression Corr.docx
Monte Carlo Simulations (UC Berkeley School of Information; July 11, 2019)
Ad

Recently uploaded (20)

PDF
Formation of Supersonic Turbulence in the Primordial Star-forming Cloud
PPTX
Microbiology with diagram medical studies .pptx
PPTX
EPIDURAL ANESTHESIA ANATOMY AND PHYSIOLOGY.pptx
PDF
Phytochemical Investigation of Miliusa longipes.pdf
PDF
Assessment of environmental effects of quarrying in Kitengela subcountyof Kaj...
PPTX
2. Earth - The Living Planet Module 2ELS
PPTX
Introduction to Fisheries Biotechnology_Lesson 1.pptx
PPTX
Introduction to Cardiovascular system_structure and functions-1
PPT
The World of Physical Science, • Labs: Safety Simulation, Measurement Practice
PDF
. Radiology Case Scenariosssssssssssssss
PPTX
7. General Toxicologyfor clinical phrmacy.pptx
PDF
Warm, water-depleted rocky exoplanets with surfaceionic liquids: A proposed c...
PDF
Unveiling a 36 billion solar mass black hole at the centre of the Cosmic Hors...
PDF
CAPERS-LRD-z9:AGas-enshroudedLittleRedDotHostingaBroad-lineActive GalacticNuc...
PDF
Looking into the jet cone of the neutrino-associated very high-energy blazar ...
PDF
HPLC-PPT.docx high performance liquid chromatography
PPTX
INTRODUCTION TO EVS | Concept of sustainability
DOCX
Q1_LE_Mathematics 8_Lesson 5_Week 5.docx
PPTX
ognitive-behavioral therapy, mindfulness-based approaches, coping skills trai...
PPTX
Taita Taveta Laboratory Technician Workshop Presentation.pptx
Formation of Supersonic Turbulence in the Primordial Star-forming Cloud
Microbiology with diagram medical studies .pptx
EPIDURAL ANESTHESIA ANATOMY AND PHYSIOLOGY.pptx
Phytochemical Investigation of Miliusa longipes.pdf
Assessment of environmental effects of quarrying in Kitengela subcountyof Kaj...
2. Earth - The Living Planet Module 2ELS
Introduction to Fisheries Biotechnology_Lesson 1.pptx
Introduction to Cardiovascular system_structure and functions-1
The World of Physical Science, • Labs: Safety Simulation, Measurement Practice
. Radiology Case Scenariosssssssssssssss
7. General Toxicologyfor clinical phrmacy.pptx
Warm, water-depleted rocky exoplanets with surfaceionic liquids: A proposed c...
Unveiling a 36 billion solar mass black hole at the centre of the Cosmic Hors...
CAPERS-LRD-z9:AGas-enshroudedLittleRedDotHostingaBroad-lineActive GalacticNuc...
Looking into the jet cone of the neutrino-associated very high-energy blazar ...
HPLC-PPT.docx high performance liquid chromatography
INTRODUCTION TO EVS | Concept of sustainability
Q1_LE_Mathematics 8_Lesson 5_Week 5.docx
ognitive-behavioral therapy, mindfulness-based approaches, coping skills trai...
Taita Taveta Laboratory Technician Workshop Presentation.pptx
Ad

Inversion_Parmetrization_under_det_problem.pdf

  • 1. Slides from Ved Lekic INVERSE THEORY 1
  • 2. 2
  • 3. Parameterization u What quantities (parameters) should we consider? u Consider pairs of quantities, such as measurements (y) made at conditions (x) Measurements (data) Conditions (independent variable)
  • 4. Parameterization u Physical model may provide insight into how to parameterize a problem u Sometimes we seek “simplest” model (fewest parameters) that can fit data Measurements (data) Conditions (independent variable)
  • 5. Law of parsimony – a useful heuristic u Preference for “simpler” models is often justified by invoking “Occam’s Razor”. u e.g. “It is vain to do with more what you can do with less.” – some 13th century monk u Karl Popper argues that simplest models are preferable because they can be more easily falsified u "I remember my friend Johnny von Neumann used to say, with four parameters I can fit an elephant, and with five I can make him wiggle his trunk.“ -- Enrico Fermi, Nature (2004). Measurements (data) Conditions (independent variable)
  • 6. Under-determined problem u With 1 data point, we cannot uniquely determine slope and y- intercept u Complete trade-off between m1 and m2! Measurements (data) Conditions (independent variable)
  • 7. Even-determined problem u When we have 2 data points, we can uniquely determine m1 and m2 Measurements (data) Conditions (independent variable)
  • 8. Over-determined problem u With >2 points, the line is not guaranteed to pass through all points. u What do we choose? u What’s the BEST line … objectively speaking? Measurements (data) Conditions (independent variable)
  • 9. The “objective” function u We introduce an objective (a.k.a. cost, misfit) function that is minimum when our line provides the “best” fit u i.e. we minimize εj’s Measurements (data) Conditions (independent variable) ε4 ε3 ε2 ε1
  • 10. Choices of Norm for objective functions u L-1 u Sum of absolute values of residuals u Tends to ignore outliers u L-2 u Sum of squared residuals u Tries to fit outliers u “Least-squares” u L-∞ u Maximum residual u Fit is “never terrible” but usually not good, either u “Minimax” φ 𝑚 = $ !"# $! 𝑑! − 𝑔!(𝑚) φ 𝑚 = 𝑚𝑎𝑥 𝑑! − 𝑔!(𝑚) φ 𝑚 = $ !"# $! 𝑑! − 𝑔!(𝑚) %
  • 11. What about data uncertainty? u If error of one measurement is expected to be larger than that of another, then the line need not pass close to it, nor should the objective function count it as much as more accurate measurements Measurements (data) Conditions (independent variable)
  • 12. What about data uncertainty? u If data error is normally distributed with variance σj 2 then the correct objective function is L-2 u Each squared residual gets weighted by variance: Measurements (data) Conditions (independent variable) φ 𝑚 = $ !"# $! 𝑑! − 𝑔!(𝑚) % 𝜎! %
  • 13. u Matrix notation: Fitting a line to some points u Linear problem: u 𝑦! = 𝑚"𝑥! + 𝑚# Y - Measurements (data) X - Conditions (independent variable) φ 𝑚 = $ !"# $! 𝑑! − 𝑔!(𝑚) % 𝜎! % 𝑥# 1 𝑥% 1 𝑥& 1 𝑥' 1 𝑚! 𝑚" 𝑦! 𝑦" 𝑦# 𝑦$ = 𝐶! "#/% 𝑑 = 𝐶! "#/% 𝐺𝑚 ⁄ # (" 0 0 ⁄ # (# 0 0 0 0 0 0 0 0 ⁄ # ($ 0 0 ⁄ # (% 𝜑 = (𝑑 − 𝐺𝑚)& 𝐶! "# (𝑑 − 𝐺𝑚) ⁄ # (" 0 0 ⁄ # (# 0 0 0 0 0 0 0 0 ⁄ # ($ 0 0 ⁄ # (%
  • 14. u Matrix notation: u Minimize φ with respect to m u Condition is satisfied when: u “Least squares” estimate of model parameters: mest Fitting a line to some points 𝐶! "#/% 𝑑 = 𝐶! "#/% 𝐺𝑚 𝜑 = (𝑑 − 𝐺𝑚)& 𝐶! "# (𝑑 − 𝐺𝑚) '( ') = 0 𝑚*+, = (𝐺& 𝐶! "# 𝐺)"# 𝐺& 𝐶! "# 𝑑 u Linear problem: u 𝑦! = 𝑚"𝑥! + 𝑚# Y - Measurements (data) X - Conditions (independent variable) φ 𝑚 = $ !"# $! 𝑑! − 𝑔!(𝑚) % 𝜎! %
  • 15. How confident are we in mest? u We can evaluate the uncertainty of model parameters using the posterior model covariance matrix: " 𝐶0 u But first, let’s build some intuition… Conditions Measurements Which scenario – left or right – will yield a less uncertain estimate of slope? Measurements Conditions
  • 16. How confident are we in mest? u We can evaluate the uncertainty of model parameters using the posterior model covariance matrix: . 𝐶) u But first, let’s build some intuition… Conditions Measurements Which scenario – left or right – will yield a less uncertain estimate of slope? Measurements Conditions BETTER! à Objective function is more affected by slope variations when data uncertainty is small à more accurate data tends to yield more reliable estimates of parameters
  • 17. How confident are we in mest? u We can evaluate the uncertainty of model parameters using the posterior model covariance matrix: . 𝐶) u But first, let’s build some intuition… Conditions Measurements Which scenario – left or right – will yield a less uncertain estimate of slope? Measurements Conditions BETTER! à Predictions of widely spaced (in x) points are more sensitive to slope variations à estimates of parameters to which you are more sensitive tend to be better!
  • 18. How confident are we in mest? u We can evaluate the uncertainty of model parameters using the posterior model covariance matrix: . 𝐶) 𝑚*+, = (𝐺& 𝐶! "# 𝐺)"# 𝐺& 𝐶! "# 𝑑 " 𝐶0= (𝐺& 𝐶! "# 𝐺)"# " 𝐶0= 𝜎+-./* % 𝜎+-./*𝜎01,𝜌 𝜎+-./*𝜎01,𝜌 𝜎01, % Variance of slope estimate Variance of y- intercept estimate Correlation between estimates of slope and y-intercept à how much parameters trade-off!
  • 19. Limitation of ! 𝐶! u Posterior model covariance matrix only describes the generalized Gaussian distribution around the optimal model 𝑚*+, = (𝐺& 𝐶! "# 𝐺)"# 𝐺& 𝐶! "# 𝑑 " 𝐶0= (𝐺& 𝐶! "# 𝐺)"# " 𝐶0= 𝜎+-./* % 𝜎+-./*𝜎01,𝜌 𝜎+-./*𝜎01,𝜌 𝜎01, % Variance of slope estimate Variance of y- intercept estimate Correlation between estimates of slope and y-intercept à how much parameters trade-off!
  • 20. Schematic representation data model LINEAR PROBLEM d = Gm m est after Tarantola, 2005 Low High Low Likelihood
  • 21. Types of inverse problems Linear – least squares works and ! 𝐶) is accurate Linearizable around starting model ! 𝐶) is probably OK Tarantola, 2005
  • 22. Non-linear à multi-modal posterior on m, ! 𝐶) is woefully invalid! • Don’t use least-squares! SAMPLING!J Linearizable around most likely model • Must iterate! • ! 𝐶) might be OK Tarantola, 2005
  • 23. nothing will work! 𝐺𝑜𝑜𝑑 𝑙𝑢𝑐𝑘! Tarantola, 2005
  • 24. data model d = Gm Perfect Inverse Problem m est u Unrealistic illustration: Perfect data, perfectly known G, and all model parameters reconstructed perfectly.
  • 25. Uncertainty about relationship between data and model data model d = Gm Modeling Uncertainty Model uncertainty!
  • 27. Effect of forward modeling error u Consequences of incorrectly computing g(m) can be significant: u Equivalent to errors to data thereby increasing model estimate uncertainty Lekic and Romanowicz, 2011 u Can systematically bias the model estimate " 𝐶0= (𝐺& 𝐶! + 𝐶& "# 𝐺)"#
  • 28. Common complication… u Often, real-world problems are mixed-determined u Some parameters are well constrained u Other parameters are poorly or completely un-constrained u Sometimes it happens that det(𝐺1 𝐶2 34 𝐺) = 0 and therefore (𝐺1𝐶2 34𝐺))# – a.k.a. G-g – does NOT exist (i.e. there are an insufficient number of linearly independent constraints to determine the unknowns) u What do we do in that case?