SlideShare a Scribd company logo
International Journal of Mathematics and Statistics Invention (IJMSI)
ISSN (Online): 2321 – 4767, ISSN (Print): 2321 – 4759
www.ijmsi.org ||Volume 5 Issue 6 ||June. 2017 || PP.06-09
www.ijmsi.org 6 | Page
A Novel Bayes Factor for Inverse Model Selection Problem based
on Inverse Reference Distribution
Debashis Chatterjee 1
1
(Interdisciplinary Statistical Research Unit / Indian Statistical Institute, India)
Abstract : Statistical model selection problem can be divided into two broad categories based on Forward and
Inverse problem. Compared to a wealthy of literature available for Forward model selection, there are very few
methods applicable for Inverse model selection context. In this article we propose a novel Bayes factor for
model selection in Bayesian Inverse Problem context. The proposed Bayes Factor is specially designed for
Inverse problem with the help of Inverse Reference Distribution (IRD). We will discuss our proposal from
decision theoretic perspective.
Keywords: Bayes Factor, Model selection, Inverse problem, Inverse Reference Distribution
I. INTRODUCTION
The question of choice of model is one of the central statistical problems. The main purpose of forward
model selection problem is to infer about the model that actually generated the data, where as the inverse model
selection primarily deals with the task of finding the most suitable model which has highest likelihood to
generate the actual unobserved covariate or the unknown data generating function. Many approaches have been
proposed in literature for proceeding with this central issue in the forward sense. The main objective of Inverse
model selection is to infer about the “best" model so that one can infer about the unknown function which
generates the data. Both frequentest and Bayesian schools have plenty of methods for model selection in
forward problem. Broadly the methods fall into two categories (refer to [1]). One contains algorithms for
picking a useful model (for example : exhaustive search, stepwise selection, backward and forward selection
procedures, F tests for nested models etc.), other category contains criteria for judging the quality of a model
(for example : AIC, BIC, Bayes Factor, cross validation etc.) There exists several methods in the statistical
literature for model assessment and model selection for forward problem which purport themselves specifically
as Bayesian predictive methods but the same for Inverse model problem is scarce. Indeed, Inverse model
selection procedure available so far is not adequate in general in contrast to the wide range of practical necessity
and applicability of statistical inverse problems.
II. INVERSE PROBLEM
Suppose one is interested in inferring about covariate Ax from the given noisy observed response
By , where A and B are Hilbert space. Let the relationship between x and y is governed by following
equation (1) :
)(= xFy (1)
where BA :F is the forward operator from A to B and B is the noise associated with the
observation. In general, we allow F to be non-injective so that the standard approximate inversion of F may
not serve the purpose. Moreover, except some statistical properties,  is unknown and there is no covenant that
noisy observation y should fall in the image set of F . As stated in [2], for finite dimensional version of the
above problem there might be situations when the number of equations is smaller than the number of unknowns.
Modern statistical research is increasingly coming across the statistical inverse problems termed as “ill-posed"
where the maximum likelihood solution or classical least square may not be uniquely defined at all and with
very bad perturbation sensitivity of the classical solution.
III. NOTATIONS & PRELIMINARIES
We will use the notation )( to denote any suitable probability density without loss of generality.
From decision theoretic perspective, let the unknown state of the world is given by the parameter  .
Decision to an inference problem will be denoted by Aa equipped with suitable utility function ),( au .
For given data D and model M , we will denote beliefs about the unknown state of the world are described by
A Novel Bayes Factor for Inverse Model Selection Problem based on Inverse Reference Distribution
www.ijmsi.org 7 | Page
the posterior distribution ).,|( MD Assume 0
 to be the prior measure with Lebesgue density )(0
x for
random variable X corresponding to the outcome Ax . We will assume joint distribution for ),( XY and
the conditional random variable YX | are well defined. YX | will be denoted by Y
X . Let us assume 
comes from distribution 0
Q where 0
Q is a measure on B such that X . Let the Lebesgue density for 0
Q
be )( . Let also assume F to follow F
Q for some suitable measure F
Q .
IV. BAYES THEOREM FOR GENERAL INVERSE PROBLEM
We can generalize the Bayes theorem for inverse problem, which is much similar to [3] and [2] under the
following assumption (1).
Assumption 1 Let   0>)()(: 0
xddQxFyZ F
  
Theorem 1 (Bayes Theorem for General Inverse Problem) If 0
QQ X
 for X a.s 0
 , then under
assumption (1) the random variable Y
XYX :=| is well defined with Lebesgue density )( xy
 given by :
   )()(
1
=)()(
1
=)( 00
xdQxFy
Z
xy
Z
x Fxy
   (2)
Let us define Potential ),( yx as follows:
  .)(log=)(log:=),( Fx
dQxFyyyx   (3)
We can then rewrite Theorem (1) similar to [2] as follows:
 ),(exp)(
0
yxx
dQ
dQ Y
 (4)
V. INVERSE REFERENCE DISTRIBUTION
Given a model M , a Bayesian model usually consists of a statistical model equipped with suitable
probability density  Mxy ,|,  for the observations with a prior distribution )|( M for the model
parameters (refer to [1]).
We can obtain the following conditional density:
).,|(),,|(=),|,( MxMxyMxy   (5)
Conditioning on observed data D in the posterior distribution ),|( MD for the model parameters, we can
determine the posterior predictive distribution
 dMDxMDxyMDxy ),,
~
|(),,,
~
|
~
(=),,
~
|
~
(  (6)
Suppose we have data set  
n
iii
yxD 1=
,= with a probability distribution ),|(  ii
xy with  as model
parameter. It is of interest to assess a given model M based on the corresponding unknown covariate x . In [4]
the author has proposed a bayesian model adequacy for such scenario. The method is based on Inverse
Reference Distribution (IRD). Given a model with dataset  nixy ii
,1,2,=);,(  where i
y is the observed
response and  n
xxxX ,,,= 21
 is a set of (non-random) covariates, the approach is to obtain reference
distributions based on the posterior )|
~
( YX , where X
~
is interpreted as the unobserved random vector
corresponding to the observed covariates X (refer to [4]). Let ),
~
,( XYL denote the Likelihood treating
),
~
( X as unknown with the model parameter  . Then the posterior )|
~
( YX can be obtained from equation
(7).
.),
~
,(),
~
()|
~
(  dXYLXYX  (7)
The Inverse Reference Distribution given by equation (7). For model M it is given by equation (8).
.),,
~
,(),,
~
(),|
~
(  dMXYLMXMYX   (8)
A Novel Bayes Factor for Inverse Model Selection Problem based on Inverse Reference Distribution
www.ijmsi.org 8 | Page
VI. BAYES FACTOR FOR INVERSE MODEL SELECTION
For rest of the paper we will assume M closed view for inverse model selection, where by
assumption it is possible to enumerate all possible model structures  
k
ii
M 1=
and the truth model as *
M . We
can assign prior distribution )( i
M over the model set or specify a non-parametric model with a prior
distribution on a suitable function space (refer to [1]). For a model i
M , we will denote the parameter space as
i
 with the parameter ii
 . The posterior probability of model i
M given the coordinate X is
)|( XM i
 given by
dyXYYXMXM ii
)|(),|(=)|(   (9)
We can define Bayes Factor for Inverse Model Selection as follows:
Definition 5.1 (Bayes Factor for Inverse Model Selection) Bayes Factor of Model i
M to j
M is defined as:
.
),|(
),|(
=,
j
i
ji
MYX
MYX
B


(10)
From Inverse Reference Distribution (IRD) given by equation (8) we can compute the Inverse Bayes Factor
given by equation (10). Hence we can write equation (10) as follows:
jjjjj
j
iiiii
i
j
i
ji
dMXYLMX
dMXYLMX
MYX
MYX
B




),,
~
,(),,
~
(
),,
~
,(),,
~
(
=
),|(
),|(
=,






(11)
VII. DECISION THEORETIC PERSPECTIVE
We will now discuss the applicability of our proposed Bayes Factor for Inverse Model Selection context under
two decision theoretic viewpoints : M - closed view and M - open view.
1) M - closed view
In the M-closed view it is possible to enumerate all possible model structures  
k
ii
M 1=
. We can assign prior
distribution )( i
M over the model set or specify a non-parametric model with a prior distribution on a
suitable function space (refer to [1]). Here we assume that there exists a ‘true’ model from which the data are
coming. For predictive Inverse model selection we suggest to find the utility ),( au as
.
~
),|
~
()
~
,(=),( xdMxxauau   (12)
The task is to select the model which maximizes the utility.
In our context, let us define ),|(= ii
MYXu  to be the utility of Model i
M . Then
.=
),|(
),|(
=,
j
i
j
i
ji
u
u
MYX
MYX
B


(13)
As mentioned earlier, the main objective of forward model selection problem is to the model M that actually
generated the data, where as the inverse model selection primarily deals with the task of finding the most
suitable model M which has highest likelihood to generate the actual unobserved covariate *
X or function
responsible for generating the data (outcome). Now if we can find the truth model *
M under M - closed view
then it serves the both purpose then that should be most desirable answer to both problem, known as oracle
solution. But In practical often there is no direct method to find the oracle solution. Notice that, the most
suitable forward model may not be the most suitable inverse model unless we have found a oracle (true) model.
2) M - open view
In the M-open view we can not explicitly construct the actual belief model, as there is a strong
background information about the belief that that any such model would not reflect well the properties of future
data ([1]). In other words, we cannot simply infer about a single ‘true’ model governing the observed data. As
state in [1], in this case, we assess the predictive performance of the candidate models under minimal modelling
A Novel Bayes Factor for Inverse Model Selection Problem based on Inverse Reference Distribution
www.ijmsi.org 9 | Page
assumptions. Nevertheless, we can obtain pseudo Monte Carlo samples from it which leads to sample re-use
methods like cross-validation (refer to [5]).
VIII. CONCLUSION
In this paper we have proposed some novel Bayes Factor specially designed for Inverse Model
Selection. It is worth mentioning that Assumption (1) may not be always true. In fact it heavily depends on
choice of prior. Cross validation based Inverse model adequacy is one of the very few alternative approach
when the assumption (1) is not valid, or in other words when posterior is likely to become improper. Future
research work of the author should address those issues.
ACKNOWLEDGEMENTS
This research work is funded by Indian Statistical Institute. The author would like to express gratitude to Prof.
Sourabh Bhattacharya, Interdisciplinary Statistical Research Unit, Indian Statistical Institute for sharing his
insight to make this research fruitful.
REFERENCES
[1]. Vehtari, Aki; Ojanen, Janne. A survey of Bayesian predictive methods for model assessment, selection and comparison. Statist.
Surv. 6 (2012), 142–228. doi:10.1214/12-SS102. http://projecteuclid.org/euclid.ssu/1356628931.
[2]. Dashti, Masoumeh; Stuart, Andrew M. The Bayesian Approach To Inverse Problems. eprint: arXiv:1302.6989, 2015.
[3]. Vollmer, S. Posterior Consistency for Bayesian Inverse Problems Through Stability and Regression Results (2013), Inverse
Problems, 29. Article number 125011.
[4]. Bhattacharya, S. . A Fully Bayesian Approach to Assessment of Model Adequacy in Inverse Problems (2013). Statistical
Methodology, 12, 71-83.
[5]. Bernardo, J. M. and Smith, A. F. M. (1994). Bayesian Theory. John Wiley & Sons. MR1274699.

More Related Content

PDF
Introduction to Supervised ML Concepts and Algorithms
PPT
Intro to Model Selection
PDF
Nbe rcausalpredictionv111 lecture2
PDF
G03405049058
PDF
When Classifier Selection meets Information Theory: A Unifying View
PDF
Big Data Analysis
PDF
The International Journal of Engineering and Science (The IJES)
PDF
Dimensionality Reduction Techniques In Response Surface Designs
Introduction to Supervised ML Concepts and Algorithms
Intro to Model Selection
Nbe rcausalpredictionv111 lecture2
G03405049058
When Classifier Selection meets Information Theory: A Unifying View
Big Data Analysis
The International Journal of Engineering and Science (The IJES)
Dimensionality Reduction Techniques In Response Surface Designs

What's hot (20)

PDF
A Class of Continuous Implicit Seventh-eight method for solving y’ = f(x, y) ...
PDF
Investigations of certain estimators for modeling panel data under violations...
PDF
Petrini - MSc Thesis
PDF
A derivative free high ordered hybrid equation solver
PDF
Transportation Problem with Pentagonal Intuitionistic Fuzzy Numbers Solved Us...
DOCX
Clustering techniques final
PPTX
Conditional mixture model for modeling attributed dyadic data
PDF
directed-research-report
PDF
Numerical Investigation of Higher Order Nonlinear Problem in the Calculus Of ...
PDF
A note on bayesian one way repeated measurements
PDF
A statistical criterion for reducing indeterminacy in linear causal modeling
PDF
11.solution of a subclass of lane emden differential equation by variational ...
PDF
Solution of a subclass of lane emden differential equation by variational ite...
PDF
Seattle.Slides.7
PDF
Solution of second kind volterra integro equations using linear
PDF
High-Dimensional Methods: Examples for Inference on Structural Effects
PDF
A Novel Algorithm for Design Tree Classification with PCA
PDF
B120512
PPTX
Handling missing data with expectation maximization algorithm
PDF
Finite Difference Method for Nonlocal Singularly Perturbed Problem
A Class of Continuous Implicit Seventh-eight method for solving y’ = f(x, y) ...
Investigations of certain estimators for modeling panel data under violations...
Petrini - MSc Thesis
A derivative free high ordered hybrid equation solver
Transportation Problem with Pentagonal Intuitionistic Fuzzy Numbers Solved Us...
Clustering techniques final
Conditional mixture model for modeling attributed dyadic data
directed-research-report
Numerical Investigation of Higher Order Nonlinear Problem in the Calculus Of ...
A note on bayesian one way repeated measurements
A statistical criterion for reducing indeterminacy in linear causal modeling
11.solution of a subclass of lane emden differential equation by variational ...
Solution of a subclass of lane emden differential equation by variational ite...
Seattle.Slides.7
Solution of second kind volterra integro equations using linear
High-Dimensional Methods: Examples for Inference on Structural Effects
A Novel Algorithm for Design Tree Classification with PCA
B120512
Handling missing data with expectation maximization algorithm
Finite Difference Method for Nonlocal Singularly Perturbed Problem
Ad

Similar to A Novel Bayes Factor for Inverse Model Selection Problem based on Inverse Reference Distribution (20)

PDF
Optimistic decision making using an
PDF
An Efficient Approach For Solving Stiff Nonlinear Boundary Value Problems
PPTX
HOMOGENOUS DEFERENTIAL EQUATIONS ,IN BIOINFOMATICS2.pptx
PDF
MEAN ABSOLUTE DEVIATION FOR HYPEREXPONENTIAL AND HYPOEXPONENTIAL DISTRIBUTION
PDF
Mean Absolute Deviation for Hyperexponential and Hypoexponential Distributions
PDF
Slides csm
PDF
MUMS: Bayesian, Fiducial, and Frequentist Conference - Inference on Treatment...
PDF
F5233444
PDF
Asymptotic properties of bayes factor in one way repeated measurements model
PDF
Asymptotic properties of bayes factor in one way repeated measurements model
PDF
Bayesian Variable Selection in Linear Regression and A Comparison
PDF
Large sample property of the bayes factor in a spline semiparametric regressi...
PPTX
unit 3 regression.pptx
PDF
A New Approach for Ranking Shadowed Fuzzy Numbers and its Application
PDF
Application NAÏVE BAYES CLASSIFIER for cancer
PDF
On Extension of Weibull Distribution with Bayesian Analysis using S-Plus Soft...
PDF
A Mathematical Programming Approach for Selection of Variables in Cluster Ana...
PPT
Elementary Probability and Information Theory
PDF
testing as a mixture estimation problem
Optimistic decision making using an
An Efficient Approach For Solving Stiff Nonlinear Boundary Value Problems
HOMOGENOUS DEFERENTIAL EQUATIONS ,IN BIOINFOMATICS2.pptx
MEAN ABSOLUTE DEVIATION FOR HYPEREXPONENTIAL AND HYPOEXPONENTIAL DISTRIBUTION
Mean Absolute Deviation for Hyperexponential and Hypoexponential Distributions
Slides csm
MUMS: Bayesian, Fiducial, and Frequentist Conference - Inference on Treatment...
F5233444
Asymptotic properties of bayes factor in one way repeated measurements model
Asymptotic properties of bayes factor in one way repeated measurements model
Bayesian Variable Selection in Linear Regression and A Comparison
Large sample property of the bayes factor in a spline semiparametric regressi...
unit 3 regression.pptx
A New Approach for Ranking Shadowed Fuzzy Numbers and its Application
Application NAÏVE BAYES CLASSIFIER for cancer
On Extension of Weibull Distribution with Bayesian Analysis using S-Plus Soft...
A Mathematical Programming Approach for Selection of Variables in Cluster Ana...
Elementary Probability and Information Theory
testing as a mixture estimation problem
Ad

Recently uploaded (20)

PDF
composite construction of structures.pdf
PPTX
Recipes for Real Time Voice AI WebRTC, SLMs and Open Source Software.pptx
PDF
keyrequirementskkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkk
PPT
Project quality management in manufacturing
PPTX
Foundation to blockchain - A guide to Blockchain Tech
PDF
Mitigating Risks through Effective Management for Enhancing Organizational Pe...
PDF
July 2025 - Top 10 Read Articles in International Journal of Software Enginee...
PPTX
CH1 Production IntroductoryConcepts.pptx
PPTX
M Tech Sem 1 Civil Engineering Environmental Sciences.pptx
PPTX
CYBER-CRIMES AND SECURITY A guide to understanding
PPTX
bas. eng. economics group 4 presentation 1.pptx
PDF
Embodied AI: Ushering in the Next Era of Intelligent Systems
PPTX
Geodesy 1.pptx...............................................
PPTX
Construction Project Organization Group 2.pptx
PDF
Digital Logic Computer Design lecture notes
PPTX
Welding lecture in detail for understanding
PDF
Mohammad Mahdi Farshadian CV - Prospective PhD Student 2026
PDF
Well-logging-methods_new................
PDF
Operating System & Kernel Study Guide-1 - converted.pdf
PDF
PRIZ Academy - 9 Windows Thinking Where to Invest Today to Win Tomorrow.pdf
composite construction of structures.pdf
Recipes for Real Time Voice AI WebRTC, SLMs and Open Source Software.pptx
keyrequirementskkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkk
Project quality management in manufacturing
Foundation to blockchain - A guide to Blockchain Tech
Mitigating Risks through Effective Management for Enhancing Organizational Pe...
July 2025 - Top 10 Read Articles in International Journal of Software Enginee...
CH1 Production IntroductoryConcepts.pptx
M Tech Sem 1 Civil Engineering Environmental Sciences.pptx
CYBER-CRIMES AND SECURITY A guide to understanding
bas. eng. economics group 4 presentation 1.pptx
Embodied AI: Ushering in the Next Era of Intelligent Systems
Geodesy 1.pptx...............................................
Construction Project Organization Group 2.pptx
Digital Logic Computer Design lecture notes
Welding lecture in detail for understanding
Mohammad Mahdi Farshadian CV - Prospective PhD Student 2026
Well-logging-methods_new................
Operating System & Kernel Study Guide-1 - converted.pdf
PRIZ Academy - 9 Windows Thinking Where to Invest Today to Win Tomorrow.pdf

A Novel Bayes Factor for Inverse Model Selection Problem based on Inverse Reference Distribution

  • 1. International Journal of Mathematics and Statistics Invention (IJMSI) ISSN (Online): 2321 – 4767, ISSN (Print): 2321 – 4759 www.ijmsi.org ||Volume 5 Issue 6 ||June. 2017 || PP.06-09 www.ijmsi.org 6 | Page A Novel Bayes Factor for Inverse Model Selection Problem based on Inverse Reference Distribution Debashis Chatterjee 1 1 (Interdisciplinary Statistical Research Unit / Indian Statistical Institute, India) Abstract : Statistical model selection problem can be divided into two broad categories based on Forward and Inverse problem. Compared to a wealthy of literature available for Forward model selection, there are very few methods applicable for Inverse model selection context. In this article we propose a novel Bayes factor for model selection in Bayesian Inverse Problem context. The proposed Bayes Factor is specially designed for Inverse problem with the help of Inverse Reference Distribution (IRD). We will discuss our proposal from decision theoretic perspective. Keywords: Bayes Factor, Model selection, Inverse problem, Inverse Reference Distribution I. INTRODUCTION The question of choice of model is one of the central statistical problems. The main purpose of forward model selection problem is to infer about the model that actually generated the data, where as the inverse model selection primarily deals with the task of finding the most suitable model which has highest likelihood to generate the actual unobserved covariate or the unknown data generating function. Many approaches have been proposed in literature for proceeding with this central issue in the forward sense. The main objective of Inverse model selection is to infer about the “best" model so that one can infer about the unknown function which generates the data. Both frequentest and Bayesian schools have plenty of methods for model selection in forward problem. Broadly the methods fall into two categories (refer to [1]). One contains algorithms for picking a useful model (for example : exhaustive search, stepwise selection, backward and forward selection procedures, F tests for nested models etc.), other category contains criteria for judging the quality of a model (for example : AIC, BIC, Bayes Factor, cross validation etc.) There exists several methods in the statistical literature for model assessment and model selection for forward problem which purport themselves specifically as Bayesian predictive methods but the same for Inverse model problem is scarce. Indeed, Inverse model selection procedure available so far is not adequate in general in contrast to the wide range of practical necessity and applicability of statistical inverse problems. II. INVERSE PROBLEM Suppose one is interested in inferring about covariate Ax from the given noisy observed response By , where A and B are Hilbert space. Let the relationship between x and y is governed by following equation (1) : )(= xFy (1) where BA :F is the forward operator from A to B and B is the noise associated with the observation. In general, we allow F to be non-injective so that the standard approximate inversion of F may not serve the purpose. Moreover, except some statistical properties,  is unknown and there is no covenant that noisy observation y should fall in the image set of F . As stated in [2], for finite dimensional version of the above problem there might be situations when the number of equations is smaller than the number of unknowns. Modern statistical research is increasingly coming across the statistical inverse problems termed as “ill-posed" where the maximum likelihood solution or classical least square may not be uniquely defined at all and with very bad perturbation sensitivity of the classical solution. III. NOTATIONS & PRELIMINARIES We will use the notation )( to denote any suitable probability density without loss of generality. From decision theoretic perspective, let the unknown state of the world is given by the parameter  . Decision to an inference problem will be denoted by Aa equipped with suitable utility function ),( au . For given data D and model M , we will denote beliefs about the unknown state of the world are described by
  • 2. A Novel Bayes Factor for Inverse Model Selection Problem based on Inverse Reference Distribution www.ijmsi.org 7 | Page the posterior distribution ).,|( MD Assume 0  to be the prior measure with Lebesgue density )(0 x for random variable X corresponding to the outcome Ax . We will assume joint distribution for ),( XY and the conditional random variable YX | are well defined. YX | will be denoted by Y X . Let us assume  comes from distribution 0 Q where 0 Q is a measure on B such that X . Let the Lebesgue density for 0 Q be )( . Let also assume F to follow F Q for some suitable measure F Q . IV. BAYES THEOREM FOR GENERAL INVERSE PROBLEM We can generalize the Bayes theorem for inverse problem, which is much similar to [3] and [2] under the following assumption (1). Assumption 1 Let   0>)()(: 0 xddQxFyZ F    Theorem 1 (Bayes Theorem for General Inverse Problem) If 0 QQ X  for X a.s 0  , then under assumption (1) the random variable Y XYX :=| is well defined with Lebesgue density )( xy  given by :    )()( 1 =)()( 1 =)( 00 xdQxFy Z xy Z x Fxy    (2) Let us define Potential ),( yx as follows:   .)(log=)(log:=),( Fx dQxFyyyx   (3) We can then rewrite Theorem (1) similar to [2] as follows:  ),(exp)( 0 yxx dQ dQ Y  (4) V. INVERSE REFERENCE DISTRIBUTION Given a model M , a Bayesian model usually consists of a statistical model equipped with suitable probability density  Mxy ,|,  for the observations with a prior distribution )|( M for the model parameters (refer to [1]). We can obtain the following conditional density: ).,|(),,|(=),|,( MxMxyMxy   (5) Conditioning on observed data D in the posterior distribution ),|( MD for the model parameters, we can determine the posterior predictive distribution  dMDxMDxyMDxy ),, ~ |(),,, ~ | ~ (=),, ~ | ~ (  (6) Suppose we have data set   n iii yxD 1= ,= with a probability distribution ),|(  ii xy with  as model parameter. It is of interest to assess a given model M based on the corresponding unknown covariate x . In [4] the author has proposed a bayesian model adequacy for such scenario. The method is based on Inverse Reference Distribution (IRD). Given a model with dataset  nixy ii ,1,2,=);,(  where i y is the observed response and  n xxxX ,,,= 21  is a set of (non-random) covariates, the approach is to obtain reference distributions based on the posterior )| ~ ( YX , where X ~ is interpreted as the unobserved random vector corresponding to the observed covariates X (refer to [4]). Let ), ~ ,( XYL denote the Likelihood treating ), ~ ( X as unknown with the model parameter  . Then the posterior )| ~ ( YX can be obtained from equation (7). .), ~ ,(), ~ ()| ~ (  dXYLXYX  (7) The Inverse Reference Distribution given by equation (7). For model M it is given by equation (8). .),, ~ ,(),, ~ (),| ~ (  dMXYLMXMYX   (8)
  • 3. A Novel Bayes Factor for Inverse Model Selection Problem based on Inverse Reference Distribution www.ijmsi.org 8 | Page VI. BAYES FACTOR FOR INVERSE MODEL SELECTION For rest of the paper we will assume M closed view for inverse model selection, where by assumption it is possible to enumerate all possible model structures   k ii M 1= and the truth model as * M . We can assign prior distribution )( i M over the model set or specify a non-parametric model with a prior distribution on a suitable function space (refer to [1]). For a model i M , we will denote the parameter space as i  with the parameter ii  . The posterior probability of model i M given the coordinate X is )|( XM i  given by dyXYYXMXM ii )|(),|(=)|(   (9) We can define Bayes Factor for Inverse Model Selection as follows: Definition 5.1 (Bayes Factor for Inverse Model Selection) Bayes Factor of Model i M to j M is defined as: . ),|( ),|( =, j i ji MYX MYX B   (10) From Inverse Reference Distribution (IRD) given by equation (8) we can compute the Inverse Bayes Factor given by equation (10). Hence we can write equation (10) as follows: jjjjj j iiiii i j i ji dMXYLMX dMXYLMX MYX MYX B     ),, ~ ,(),, ~ ( ),, ~ ,(),, ~ ( = ),|( ),|( =,       (11) VII. DECISION THEORETIC PERSPECTIVE We will now discuss the applicability of our proposed Bayes Factor for Inverse Model Selection context under two decision theoretic viewpoints : M - closed view and M - open view. 1) M - closed view In the M-closed view it is possible to enumerate all possible model structures   k ii M 1= . We can assign prior distribution )( i M over the model set or specify a non-parametric model with a prior distribution on a suitable function space (refer to [1]). Here we assume that there exists a ‘true’ model from which the data are coming. For predictive Inverse model selection we suggest to find the utility ),( au as . ~ ),| ~ () ~ ,(=),( xdMxxauau   (12) The task is to select the model which maximizes the utility. In our context, let us define ),|(= ii MYXu  to be the utility of Model i M . Then .= ),|( ),|( =, j i j i ji u u MYX MYX B   (13) As mentioned earlier, the main objective of forward model selection problem is to the model M that actually generated the data, where as the inverse model selection primarily deals with the task of finding the most suitable model M which has highest likelihood to generate the actual unobserved covariate * X or function responsible for generating the data (outcome). Now if we can find the truth model * M under M - closed view then it serves the both purpose then that should be most desirable answer to both problem, known as oracle solution. But In practical often there is no direct method to find the oracle solution. Notice that, the most suitable forward model may not be the most suitable inverse model unless we have found a oracle (true) model. 2) M - open view In the M-open view we can not explicitly construct the actual belief model, as there is a strong background information about the belief that that any such model would not reflect well the properties of future data ([1]). In other words, we cannot simply infer about a single ‘true’ model governing the observed data. As state in [1], in this case, we assess the predictive performance of the candidate models under minimal modelling
  • 4. A Novel Bayes Factor for Inverse Model Selection Problem based on Inverse Reference Distribution www.ijmsi.org 9 | Page assumptions. Nevertheless, we can obtain pseudo Monte Carlo samples from it which leads to sample re-use methods like cross-validation (refer to [5]). VIII. CONCLUSION In this paper we have proposed some novel Bayes Factor specially designed for Inverse Model Selection. It is worth mentioning that Assumption (1) may not be always true. In fact it heavily depends on choice of prior. Cross validation based Inverse model adequacy is one of the very few alternative approach when the assumption (1) is not valid, or in other words when posterior is likely to become improper. Future research work of the author should address those issues. ACKNOWLEDGEMENTS This research work is funded by Indian Statistical Institute. The author would like to express gratitude to Prof. Sourabh Bhattacharya, Interdisciplinary Statistical Research Unit, Indian Statistical Institute for sharing his insight to make this research fruitful. REFERENCES [1]. Vehtari, Aki; Ojanen, Janne. A survey of Bayesian predictive methods for model assessment, selection and comparison. Statist. Surv. 6 (2012), 142–228. doi:10.1214/12-SS102. http://projecteuclid.org/euclid.ssu/1356628931. [2]. Dashti, Masoumeh; Stuart, Andrew M. The Bayesian Approach To Inverse Problems. eprint: arXiv:1302.6989, 2015. [3]. Vollmer, S. Posterior Consistency for Bayesian Inverse Problems Through Stability and Regression Results (2013), Inverse Problems, 29. Article number 125011. [4]. Bhattacharya, S. . A Fully Bayesian Approach to Assessment of Model Adequacy in Inverse Problems (2013). Statistical Methodology, 12, 71-83. [5]. Bernardo, J. M. and Smith, A. F. M. (1994). Bayesian Theory. John Wiley & Sons. MR1274699.