SlideShare a Scribd company logo
A short and naive introduction to using network in
prediction models
Nathalie Vialaneix
nathalie.vialaneix@inra.fr
http://www.nathalievialaneix.eu
EpiFun
December 7th, 2018 - Paris
Nathalie Vialaneix | Using network in prediction models 1/22
Outline
1 A very short introduction
2 Some background on graphs and kernels
3 Using Laplacian in prediction
4 Coming back to our problem... what about SNP networks?
Nathalie Vialaneix | Using network in prediction models 2/22
What is this presentation about?
How to integrate network a priori in prediction models with application to
biology?
Nathalie Vialaneix | Using network in prediction models 3/22
Disclaimer
has been made with the LA RACHE methodology
might contain some (not understandable) maths...
is probably not exhaustive
should provide an intuitive idea of the basic
concepts and a few directions
Nathalie Vialaneix | Using network in prediction models 4/22
Disclaimer
has been made with the LA RACHE methodology
might contain some (not understandable) maths...
is probably not exhaustive
should provide an intuitive idea of the basic
concepts and a few directions
Material:
References at the end of the slides
these slides on my website
http://www.nathalievialaneix.eu/seminars2018.html
most articles available online at http://nextcloud.
nathalievilla.org/index.php/s/VLlheqpwhwD8eeZ (ask me to
be granted write rights)
Nathalie Vialaneix | Using network in prediction models 4/22
Outline
1 A very short introduction
2 Some background on graphs and kernels
3 Using Laplacian in prediction
4 Coming back to our problem... what about SNP networks?
Nathalie Vialaneix | Using network in prediction models 5/22
Background and notations
What we have: a network (graph), G, with p nodes, v1, . . . , vp and edges
between these nodes
Example: nodes are genes; edges are known regulatory information or co-expression
Nathalie Vialaneix | Using network in prediction models 6/22
Background and notations
What we have: a network (graph), G, with p nodes, v1, . . . , vp and edges
between these nodes
Example: nodes are genes; edges are known regulatory information or co-expression
An important matrix: the Laplacian
LG
ij
=









−1 if i , j and vi and vj are linked by an edge
0 if i , j and vi and vj are not linked by an edge
di if i = j
with di the degree (i.e., the number of edges) of node vi.
Minor note (however important for those who like linear algebra): rows and columns
of this matrix sum to 0. It is equivalent to notice that 0 is an eigenvalue (the smallest) with
1p its eigenvector.
Nathalie Vialaneix | Using network in prediction models 6/22
Some properties of the Laplacian
Relations with the graph structure:
1
2
3
4
5
has a null space
(eigenvalues = 0) spanned by the eigenvectors





















1
1
1
0
0





















and





















0
0
0
1
1





















(similar to
the fact that the vector 1p spans the null space for connected graphs).
Nathalie Vialaneix | Using network in prediction models 7/22
Some properties of the Laplacian
Relations with the graph structure:
Random walk point of view: If we consider a random walk on the graph
with probability to jump from one node to the other equal to 1
di
then the
average time to go from one node to another (commute time) is given by
L+ [Fouss et al., 2007] (commute time kernel).
Diffusion process point of view: If we consider a (heat) diffusion process
on the graph where each node “send” a fraction β > 0 of its value (heat)
through the edges, exp−βL
is related to the covariance of this diffusion
process [Kondor and Lafferty, 2002] (heat kernel).
Nathalie Vialaneix | Using network in prediction models 7/22
Relation between RatioCut and Laplacian (graph
clustering)
[von Luxburg, 2007] shows that find a partition of the graph into two clusters
that minimize:
RatioCut(C1, C2) =
1
2
2
X
k=1
X
vi∈Ck , vj<Ck vi∼vj
1
|Ck |
is equivalent to the following constrained problem:
min
C1, ,C2
v>
Lv st v ⊥ 1n and kvk =
√
n
for v the vector of Rn
obtained from the partition by:
vi =
( p
(|C2|)/|C1| if vi ∈ C1
−
p
|C1|/(|C2|) otherwise.
Nathalie Vialaneix | Using network in prediction models 8/22
Eigendecomposition of the Laplacian
L is symmetric and positive (not definite positive) so it can be decomposed
into:
L =
p
X
i=1
λieie>
i
with λi the eigenvalues (in increasing order) and ei the orthonormal
eigenvectors in Rp
.
Nathalie Vialaneix | Using network in prediction models 9/22
Eigendecomposition of the Laplacian
L is symmetric and positive (not definite positive) so it can be decomposed
into:
L =
p
X
i=1
λieie>
i
with λi the eigenvalues (in increasing order) and ei the orthonormal
eigenvectors in Rp
.
If you want to extract the most relevant information from the network, use
the smallest eigenvalues with:
low pass filter (similar to signal processing): FG =
Pr
i=1 λieie>
i
for
r < p
regularization FG =
Pp
i=1
φ(λi)eie>
i
with φ(x) = e−βλi or 1
λi
for
instance (is, most of the time, a kernel)
Nathalie Vialaneix | Using network in prediction models 9/22
Take home messages
eigenvectors of the Laplacian that are associated to the smallest
eigenvalues are strongly related to the graph structure
many kernels have been derived from the Laplacian
[Smola and Kondor, 2003] that perform regularization on graphs that
can be used to measure similarities between nodes of the graph
Nathalie Vialaneix | Using network in prediction models 10/22
Outline
1 A very short introduction
2 Some background on graphs and kernels
3 Using Laplacian in prediction
4 Coming back to our problem... what about SNP networks?
Nathalie Vialaneix | Using network in prediction models 11/22
Background
Problem: predict Y (numerical) from X (multivariate, dimension p) with a
linear model:
y
|{z}
vector, length n
= X
|{z}
matrix, dimension n×p
× β
|{z}
vector to be estimated, length p
+
Nathalie Vialaneix | Using network in prediction models 12/22
Background
Problem: predict Y (numerical) from X (multivariate, dimension p) with a
linear model:
y
|{z}
vector, length n
= X
|{z}
matrix, dimension n×p
× β
|{z}
vector to be estimated, length p
+
Examples:
[Rapaport et al., 2007]: Y is “Radiated/Not radiated sample” and X is
gene expression. A network is given on the p genes based on KEGG
metabolic pathways
[Li and Li, 2008]: Y is time to death (Glioblastoma) and X is gene
expression. A network is given on the p genes based on KEGG
metabolic pathways
Nathalie Vialaneix | Using network in prediction models 12/22
Sketch of main directions
1 use a kernel based on the Laplacian and its associated dot product to
compute a distance
2 use a standard linear model but regularize/penalize it with the
Laplacian norm
Nathalie Vialaneix | Using network in prediction models 13/22
Working in the feature space...
[Rapaport et al., 2007]
xi is gene expression (length p) for sample i
Standard approach
→ a gene expression profile:
xi =
p
X
j=1
xij

























0
.
.
.
1
.
.
.
0

























→ dot product is:
p
X
j=1
(xijxi0j)
Nathalie Vialaneix | Using network in prediction models 14/22
Working in the feature space...
[Rapaport et al., 2007]
xi is gene expression (length p) for sample i
Standard approach
→ a gene expression profile:
xi =
p
X
j=1
xij

























0
.
.
.
1
.
.
.
0

























→ dot product is:
p
X
j=1
(xijxi0j)
Laplacian approach
→ a gene expression profile:
Φ(xi) =
p
X
j=1
φ(λj)x̃ijej
with x̃ij = e
j xi (similar to Fourier
transform)
→ dot product is:
p
X
j=1
x
i Kφxi0
with Kφ the kernel obtained from L with φ
Nathalie Vialaneix | Using network in prediction models 14/22
Working in the feature space... and then?
[Rapaport et al., 2007]
∼ Standard linear regression with the Laplacian dot product + `2 penalty
with the Laplacian kernel norm:
arg min
w∈Rp
n
X
i=1

w
Φ(xi) − yi
2
+Ckwk2
⇔ arg min
β∈Rp
n
X
i=1

β
xi − yi
2
+Cβ
K−1
φ β
this is called a SVM...!
redundant information from two genes that are strongly connected on
the graph is taken into account: this approach enforces similar
contributions of these genes for the prediction
improved interpretation
Nathalie Vialaneix | Using network in prediction models 15/22
Direct penalty with L
[Li and Li, 2008]
arg min
β∈Rp
n
X
i=1

β
xi − yi
2
+ Cβ
Lβ +C0
kβk1
| {z }
to enforce sparsity
Nathalie Vialaneix | Using network in prediction models 16/22
Direct penalty with L
[Li and Li, 2008]
arg min
β∈Rp
n
X
i=1

β
xi − yi
2
+ Cβ
Lβ +C0
kβk1
| {z }
to enforce sparsity
Remarks:
this is a very similar idea than previously (identical penalty for
φ(·) = 1/·).
Nathalie Vialaneix | Using network in prediction models 16/22
Direct penalty with L
[Li and Li, 2008]
arg min
β∈Rp
n
X
i=1

β
xi − yi
2
+ Cβ
Lβ +C0
kβk1
| {z }
to enforce sparsity
Remarks:
this is a very similar idea than previously (identical penalty for
φ(·) = 1/·).
[Li and Li, 2010] explain that this forces connected vertices to have
similar contributions whereas in some cases, they can have opposite
contributions
⇒ use of L
β
jj0 =









−sign(βj)sign(βj0 ) if j , j0 and vj and vj0 are linked by an edge
0 if j , j0 and vj and vj0 are not linked by an edge
dj if j = j0
Nathalie Vialaneix | Using network in prediction models 16/22
Outline
1 A very short introduction
2 Some background on graphs and kernels
3 Using Laplacian in prediction
4 Coming back to our problem... what about SNP networks?
Nathalie Vialaneix | Using network in prediction models 17/22
Similar approach... with SNPs
[Azencott et al., 2013]
Background: X are SNP values, Y is a phenotype but it is not used as but
to derive a association score between each single SNP and the
phenotype, c ∈ Rp
(ex: SKAT kernel)
In addition, a network is given between SNPs (in the paper, derived from
KEGG gene network or co-expression network)
Nathalie Vialaneix | Using network in prediction models 18/22
Similar approach... with SNPs
[Azencott et al., 2013]
Background: X are SNP values, Y is a phenotype but it is not used as but
to derive a association score between each single SNP and the
phenotype, c ∈ Rp
(ex: SKAT kernel)
In addition, a network is given between SNPs (in the paper, derived from
KEGG gene network or co-expression network)
Purpose: discover sets of genetic loci that are maximally associated with a
phenotype while being connected in the underlying network
→ search for s ∈ {0, 1}p
with the association score associated to s given
by
Pp
j=1
sjcj = sc (to be maximized)
Nathalie Vialaneix | Using network in prediction models 18/22
Similar approach... with SNPs
[Azencott et al., 2013]
Background: X are SNP values, Y is a phenotype but it is not used as but
to derive a association score between each single SNP and the
phenotype, c ∈ Rp
(ex: SKAT kernel)
In addition, a network is given between SNPs (in the paper, derived from
KEGG gene network or co-expression network)
Purpose: discover sets of genetic loci that are maximally associated with a
phenotype while being connected in the underlying network
→ search for s ∈ {0, 1}p
with the association score associated to s given
by
Pp
j=1
sjcj = sc (to be maximized)
Solution:
arg max
s∈{0,1}p
s
c − Cs
Ls −C0
ksk0
| {z }
to enforce sparsity
Nathalie Vialaneix | Using network in prediction models 18/22
Another similar reference... with SNPs
[Chiquet et al., 2016]
Background: X are SNP values, Y are several phenotypes (multi-trait
prediction ⇒ multivariate regression model)
In addition, the authors do not directly use a network but a similarity that is
related to the genomic distance, L
Nathalie Vialaneix | Using network in prediction models 19/22
Another similar reference... with SNPs
[Chiquet et al., 2016]
Background: X are SNP values, Y are several phenotypes (multi-trait
prediction ⇒ multivariate regression model)
In addition, the authors do not directly use a network but a similarity that is
related to the genomic distance, L
Purpose: prediction, estimation of coefficients (in linear models) and of
direct effects are obtained using a model with i/ `2 regularization with the
norm induced by L and 2/ `1 penalization to enforce sparsity (e.g., select
SNPs)
Nathalie Vialaneix | Using network in prediction models 19/22
Other references on SNP networks
[Hu et al., 2011] their aim is SNP network construction in association
with a phenotype (for further mining): pairwise interactions detected
with entropy gain (cancer)
[Kogelman and Kadarmideen, 2014] their aim is also SNP network
construction in association with a phenotype (for clustering + mining):
combine genomic correlation (similar to LD) with genetic association
(interaction tests) in a very crude way (almost just a sum) (pig carcass
weight)
Nathalie Vialaneix | Using network in prediction models 20/22
Promising questions
using a priori SNP network in association models
comparing SNP networks (as described in the previous slide) with a
priori networks to check if this solution is potentially interesting
... that will be for the next meeting!
Nathalie Vialaneix | Using network in prediction models 21/22
Nathalie Vialaneix | Using network in prediction models 22/22
References
Azencott, C.-A., Grimm, D., Sugiyama, M., Kawahara, Y., and Borgwardt, K. M. (2013).
Efficient network-guided multi-locus association mapping with graph cuts.
Bioinformatics, 29(13):i171–i179.
Chiquet, J., Mary-Huard, T., and Robin, S. (2016).
Structured regularization for conditional Gaussian graphical models.
Statistics and Computing, pages 789–804.
Fouss, F., Pirotte, A., Renders, J., and Saerens, M. (2007).
Random-walk computation of similarities between nodes of a graph, with application to collaborative recommendation.
IEEE Transactions on Knowledge and Data Engineering, 19(3):355–369.
Hu, T., Sinnott-Armstrong, N. A., Kiralis, J. W., Andrew, A. S., Karagas, M. R., and Moore, J. H. (2011).
Characterizing genetic interactions in human disease association studies using statistical epistasis networks.
BMC Bioinformatics, 12:364.
Kogelman, L. J. and Kadarmideen, H. N. (2014).
Weighted interaction SNP hub (WISH) network method for building genetic networks for complex diseases traits using whole
genome genotype data.
BMC Systems Biology, 8(Suppl 2).
Kondor, R. and Lafferty, J. (2002).
Diffusion kernels on graphs and other discrete structures.
In Sammut, C. and Hoffmann, A., editors, Proceedings of the 19th International Conference on Machine Learning, pages
315–322, Sydney, Australia. Morgan Kaufmann Publishers Inc. San Francisco, CA, USA.
Li, C. and Li, H. (2008).
Network-constrained regularization and variable selection for analysis of genomic data.
Bioinformatics, 24(9):1175–1182.
Li, C. and Li, H. (2010).
Variable selection and regression analysis for graph-structured covariates with an application to genomics.
The Annals of Applied Statistics, 4(3):1498–1516.
Nathalie Vialaneix | Using network in prediction models 22/22
Rapaport, F., Zinovyev, A., Dutreix, M., Barillot, E., and Vert, J. (2007).
Classification of microarray data using gene networks.
BMC Bioinformatics, 8:35.
Smola, A. and Kondor, R. (2003).
Kernels and regularization on graphs.
In Warmuth, M. and Schölkopf, B., editors, Proceedings of the Conference on Learning Theory (COLT) and Kernel Workshop,
Lecture Notes in Computer Science, pages 144–158, Washington, DC, USA. Springer-Verlag Berlin Heidelberg.
von Luxburg, U. (2007).
A tutorial on spectral clustering.
Statistics and Computing, 17(4):395–416.
Nathalie Vialaneix | Using network in prediction models 22/22

More Related Content

PDF
From RNN to neural networks for cyclic undirected graphs
PDF
Graph Neural Network in practice
PDF
Convolutional networks and graph networks through kernels
PDF
Differential analyses of structures in HiC data
PDF
A review on structure learning in GNN
PDF
Kernel methods and variable selection for exploratory analysis and multi-omic...
PDF
Graph Neural Network for Phenotype Prediction
PDF
About functional SIR
From RNN to neural networks for cyclic undirected graphs
Graph Neural Network in practice
Convolutional networks and graph networks through kernels
Differential analyses of structures in HiC data
A review on structure learning in GNN
Kernel methods and variable selection for exploratory analysis and multi-omic...
Graph Neural Network for Phenotype Prediction
About functional SIR

What's hot (20)

PDF
Kernel methods for data integration in systems biology
PDF
Reproducibility and differential analysis with selfish
PDF
Methods of Manifold Learning for Dimension Reduction of Large Data Sets
PDF
Random Forest for Big Data
PDF
Kernel methods for data integration in systems biology
PDF
Investigating the 3D structure of the genome with Hi-C data analysis
PDF
Training and Inference for Deep Gaussian Processes
PDF
Gtti 10032021
PDF
Macrocanonical models for texture synthesis
PDF
010_20160216_Variational Gaussian Process
PDF
Quantitative Propagation of Chaos for SGD in Wide Neural Networks
PDF
An Importance Sampling Approach to Integrate Expert Knowledge When Learning B...
PDF
Iclr2016 vaeまとめ
PDF
Learning from (dis)similarity data
PDF
A discussion on sampling graphs to approximate network classification functions
PDF
Estimating Space-Time Covariance from Finite Sample Sets
PDF
Information Content of Complex Networks
PDF
Polynomial Matrix Decompositions
PDF
Deep Learning Opening Workshop - Horseshoe Regularization for Machine Learnin...
PDF
Kernel methods for data integration in systems biology
Reproducibility and differential analysis with selfish
Methods of Manifold Learning for Dimension Reduction of Large Data Sets
Random Forest for Big Data
Kernel methods for data integration in systems biology
Investigating the 3D structure of the genome with Hi-C data analysis
Training and Inference for Deep Gaussian Processes
Gtti 10032021
Macrocanonical models for texture synthesis
010_20160216_Variational Gaussian Process
Quantitative Propagation of Chaos for SGD in Wide Neural Networks
An Importance Sampling Approach to Integrate Expert Knowledge When Learning B...
Iclr2016 vaeまとめ
Learning from (dis)similarity data
A discussion on sampling graphs to approximate network classification functions
Estimating Space-Time Covariance from Finite Sample Sets
Information Content of Complex Networks
Polynomial Matrix Decompositions
Deep Learning Opening Workshop - Horseshoe Regularization for Machine Learnin...
Ad

Similar to A short and naive introduction to using network in prediction models (20)

PDF
Network analysis for computational biology
PDF
08 Exponential Random Graph Models (ERGM)
PDF
08 Exponential Random Graph Models (2016)
PPT
Cornell Pbsb 20090126 Nets
PDF
Sparse inverse covariance estimation using skggm
PDF
Topics In Theoretical Computer Science An Algorithmists Toolkit Lecture Notes...
PPTX
WIDS 2021--An Introduction to Network Science
PDF
Application of three graph Laplacian based semisupervised learning methods to...
PDF
Networks, Deep Learning (and COVID-19)
PDF
Quelques résultats préliminaires de l'évaluation de méthodes d'inférence de r...
PDF
Lausanne 2019 #4
PDF
Learning the structure of Gaussian Graphical models with unobserved variables...
PDF
Graph Signal Processing for Machine Learning A Review and New Perspectives - ...
PDF
Webinar on Graph Neural Networks
PDF
Application of Bayesian and Sparse Network Models for Assessing Linkage Diseq...
PDF
Grl book
PDF
Sparsenet
PDF
Regulation Analysis using Restricted Boltzmann Machines
PDF
08 Inference for Networks – DYAD Model Overview (2017)
PDF
Consensual gene co-expression network inference with multiple samples
Network analysis for computational biology
08 Exponential Random Graph Models (ERGM)
08 Exponential Random Graph Models (2016)
Cornell Pbsb 20090126 Nets
Sparse inverse covariance estimation using skggm
Topics In Theoretical Computer Science An Algorithmists Toolkit Lecture Notes...
WIDS 2021--An Introduction to Network Science
Application of three graph Laplacian based semisupervised learning methods to...
Networks, Deep Learning (and COVID-19)
Quelques résultats préliminaires de l'évaluation de méthodes d'inférence de r...
Lausanne 2019 #4
Learning the structure of Gaussian Graphical models with unobserved variables...
Graph Signal Processing for Machine Learning A Review and New Perspectives - ...
Webinar on Graph Neural Networks
Application of Bayesian and Sparse Network Models for Assessing Linkage Diseq...
Grl book
Sparsenet
Regulation Analysis using Restricted Boltzmann Machines
08 Inference for Networks – DYAD Model Overview (2017)
Consensual gene co-expression network inference with multiple samples
Ad

More from tuxette (20)

PDF
Analyse comparative de données de génomique 3D
PDF
Detecting differences between 3D genomic data: a benchmark study
PDF
Racines en haut et feuilles en bas : les arbres en maths
PDF
Méthodes à noyaux pour l’intégration de données hétérogènes
PDF
Méthodologies d'intégration de données omiques
PDF
Projets autour de l'Hi-C
PDF
Can deep learning learn chromatin structure from sequence?
PDF
Multi-omics data integration methods: kernel and other machine learning appro...
PDF
ASTERICS : une application pour intégrer des données omiques
PDF
Autour des projets Idefics et MetaboWean
PDF
Rserve, renv, flask, Vue.js dans un docker pour intégrer des données omiques ...
PDF
Apprentissage pour la biologie moléculaire et l’analyse de données omiques
PDF
Intégration de données omiques multi-échelles : méthodes à noyau et autres ap...
PDF
Journal club: Validation of cluster analysis results on validation data
PDF
Overfitting or overparametrization?
PDF
Selective inference and single-cell differential analysis
PDF
SOMbrero : un package R pour les cartes auto-organisatrices
PDF
Explanable models for time series with random forest
PDF
Présentation du projet ASTERICS
PDF
Présentation du projet ASTERICS
Analyse comparative de données de génomique 3D
Detecting differences between 3D genomic data: a benchmark study
Racines en haut et feuilles en bas : les arbres en maths
Méthodes à noyaux pour l’intégration de données hétérogènes
Méthodologies d'intégration de données omiques
Projets autour de l'Hi-C
Can deep learning learn chromatin structure from sequence?
Multi-omics data integration methods: kernel and other machine learning appro...
ASTERICS : une application pour intégrer des données omiques
Autour des projets Idefics et MetaboWean
Rserve, renv, flask, Vue.js dans un docker pour intégrer des données omiques ...
Apprentissage pour la biologie moléculaire et l’analyse de données omiques
Intégration de données omiques multi-échelles : méthodes à noyau et autres ap...
Journal club: Validation of cluster analysis results on validation data
Overfitting or overparametrization?
Selective inference and single-cell differential analysis
SOMbrero : un package R pour les cartes auto-organisatrices
Explanable models for time series with random forest
Présentation du projet ASTERICS
Présentation du projet ASTERICS

Recently uploaded (20)

PDF
CAPERS-LRD-z9:AGas-enshroudedLittleRedDotHostingaBroad-lineActive GalacticNuc...
PDF
Unveiling a 36 billion solar mass black hole at the centre of the Cosmic Hors...
PPTX
ECG_Course_Presentation د.محمد صقران ppt
PPTX
famous lake in india and its disturibution and importance
PDF
AlphaEarth Foundations and the Satellite Embedding dataset
PPTX
7. General Toxicologyfor clinical phrmacy.pptx
PPT
POSITIONING IN OPERATION THEATRE ROOM.ppt
DOCX
Viruses (History, structure and composition, classification, Bacteriophage Re...
PPTX
Taita Taveta Laboratory Technician Workshop Presentation.pptx
PPT
protein biochemistry.ppt for university classes
PPTX
Classification Systems_TAXONOMY_SCIENCE8.pptx
PPTX
2. Earth - The Living Planet earth and life
PDF
SEHH2274 Organic Chemistry Notes 1 Structure and Bonding.pdf
PPTX
neck nodes and dissection types and lymph nodes levels
PDF
An interstellar mission to test astrophysical black holes
PPTX
ANEMIA WITH LEUKOPENIA MDS 07_25.pptx htggtftgt fredrctvg
PDF
. Radiology Case Scenariosssssssssssssss
PDF
VARICELLA VACCINATION: A POTENTIAL STRATEGY FOR PREVENTING MULTIPLE SCLEROSIS
PPTX
G5Q1W8 PPT SCIENCE.pptx 2025-2026 GRADE 5
PPTX
2. Earth - The Living Planet Module 2ELS
CAPERS-LRD-z9:AGas-enshroudedLittleRedDotHostingaBroad-lineActive GalacticNuc...
Unveiling a 36 billion solar mass black hole at the centre of the Cosmic Hors...
ECG_Course_Presentation د.محمد صقران ppt
famous lake in india and its disturibution and importance
AlphaEarth Foundations and the Satellite Embedding dataset
7. General Toxicologyfor clinical phrmacy.pptx
POSITIONING IN OPERATION THEATRE ROOM.ppt
Viruses (History, structure and composition, classification, Bacteriophage Re...
Taita Taveta Laboratory Technician Workshop Presentation.pptx
protein biochemistry.ppt for university classes
Classification Systems_TAXONOMY_SCIENCE8.pptx
2. Earth - The Living Planet earth and life
SEHH2274 Organic Chemistry Notes 1 Structure and Bonding.pdf
neck nodes and dissection types and lymph nodes levels
An interstellar mission to test astrophysical black holes
ANEMIA WITH LEUKOPENIA MDS 07_25.pptx htggtftgt fredrctvg
. Radiology Case Scenariosssssssssssssss
VARICELLA VACCINATION: A POTENTIAL STRATEGY FOR PREVENTING MULTIPLE SCLEROSIS
G5Q1W8 PPT SCIENCE.pptx 2025-2026 GRADE 5
2. Earth - The Living Planet Module 2ELS

A short and naive introduction to using network in prediction models

  • 1. A short and naive introduction to using network in prediction models Nathalie Vialaneix nathalie.vialaneix@inra.fr http://www.nathalievialaneix.eu EpiFun December 7th, 2018 - Paris Nathalie Vialaneix | Using network in prediction models 1/22
  • 2. Outline 1 A very short introduction 2 Some background on graphs and kernels 3 Using Laplacian in prediction 4 Coming back to our problem... what about SNP networks? Nathalie Vialaneix | Using network in prediction models 2/22
  • 3. What is this presentation about? How to integrate network a priori in prediction models with application to biology? Nathalie Vialaneix | Using network in prediction models 3/22
  • 4. Disclaimer has been made with the LA RACHE methodology might contain some (not understandable) maths... is probably not exhaustive should provide an intuitive idea of the basic concepts and a few directions Nathalie Vialaneix | Using network in prediction models 4/22
  • 5. Disclaimer has been made with the LA RACHE methodology might contain some (not understandable) maths... is probably not exhaustive should provide an intuitive idea of the basic concepts and a few directions Material: References at the end of the slides these slides on my website http://www.nathalievialaneix.eu/seminars2018.html most articles available online at http://nextcloud. nathalievilla.org/index.php/s/VLlheqpwhwD8eeZ (ask me to be granted write rights) Nathalie Vialaneix | Using network in prediction models 4/22
  • 6. Outline 1 A very short introduction 2 Some background on graphs and kernels 3 Using Laplacian in prediction 4 Coming back to our problem... what about SNP networks? Nathalie Vialaneix | Using network in prediction models 5/22
  • 7. Background and notations What we have: a network (graph), G, with p nodes, v1, . . . , vp and edges between these nodes Example: nodes are genes; edges are known regulatory information or co-expression Nathalie Vialaneix | Using network in prediction models 6/22
  • 8. Background and notations What we have: a network (graph), G, with p nodes, v1, . . . , vp and edges between these nodes Example: nodes are genes; edges are known regulatory information or co-expression An important matrix: the Laplacian LG ij =          −1 if i , j and vi and vj are linked by an edge 0 if i , j and vi and vj are not linked by an edge di if i = j with di the degree (i.e., the number of edges) of node vi. Minor note (however important for those who like linear algebra): rows and columns of this matrix sum to 0. It is equivalent to notice that 0 is an eigenvalue (the smallest) with 1p its eigenvector. Nathalie Vialaneix | Using network in prediction models 6/22
  • 9. Some properties of the Laplacian Relations with the graph structure: 1 2 3 4 5 has a null space (eigenvalues = 0) spanned by the eigenvectors                      1 1 1 0 0                      and                      0 0 0 1 1                      (similar to the fact that the vector 1p spans the null space for connected graphs). Nathalie Vialaneix | Using network in prediction models 7/22
  • 10. Some properties of the Laplacian Relations with the graph structure: Random walk point of view: If we consider a random walk on the graph with probability to jump from one node to the other equal to 1 di then the average time to go from one node to another (commute time) is given by L+ [Fouss et al., 2007] (commute time kernel). Diffusion process point of view: If we consider a (heat) diffusion process on the graph where each node “send” a fraction β > 0 of its value (heat) through the edges, exp−βL is related to the covariance of this diffusion process [Kondor and Lafferty, 2002] (heat kernel). Nathalie Vialaneix | Using network in prediction models 7/22
  • 11. Relation between RatioCut and Laplacian (graph clustering) [von Luxburg, 2007] shows that find a partition of the graph into two clusters that minimize: RatioCut(C1, C2) = 1 2 2 X k=1 X vi∈Ck , vj<Ck vi∼vj 1 |Ck | is equivalent to the following constrained problem: min C1, ,C2 v> Lv st v ⊥ 1n and kvk = √ n for v the vector of Rn obtained from the partition by: vi = ( p (|C2|)/|C1| if vi ∈ C1 − p |C1|/(|C2|) otherwise. Nathalie Vialaneix | Using network in prediction models 8/22
  • 12. Eigendecomposition of the Laplacian L is symmetric and positive (not definite positive) so it can be decomposed into: L = p X i=1 λieie> i with λi the eigenvalues (in increasing order) and ei the orthonormal eigenvectors in Rp . Nathalie Vialaneix | Using network in prediction models 9/22
  • 13. Eigendecomposition of the Laplacian L is symmetric and positive (not definite positive) so it can be decomposed into: L = p X i=1 λieie> i with λi the eigenvalues (in increasing order) and ei the orthonormal eigenvectors in Rp . If you want to extract the most relevant information from the network, use the smallest eigenvalues with: low pass filter (similar to signal processing): FG = Pr i=1 λieie> i for r < p regularization FG = Pp i=1 φ(λi)eie> i with φ(x) = e−βλi or 1 λi for instance (is, most of the time, a kernel) Nathalie Vialaneix | Using network in prediction models 9/22
  • 14. Take home messages eigenvectors of the Laplacian that are associated to the smallest eigenvalues are strongly related to the graph structure many kernels have been derived from the Laplacian [Smola and Kondor, 2003] that perform regularization on graphs that can be used to measure similarities between nodes of the graph Nathalie Vialaneix | Using network in prediction models 10/22
  • 15. Outline 1 A very short introduction 2 Some background on graphs and kernels 3 Using Laplacian in prediction 4 Coming back to our problem... what about SNP networks? Nathalie Vialaneix | Using network in prediction models 11/22
  • 16. Background Problem: predict Y (numerical) from X (multivariate, dimension p) with a linear model: y |{z} vector, length n = X |{z} matrix, dimension n×p × β |{z} vector to be estimated, length p + Nathalie Vialaneix | Using network in prediction models 12/22
  • 17. Background Problem: predict Y (numerical) from X (multivariate, dimension p) with a linear model: y |{z} vector, length n = X |{z} matrix, dimension n×p × β |{z} vector to be estimated, length p + Examples: [Rapaport et al., 2007]: Y is “Radiated/Not radiated sample” and X is gene expression. A network is given on the p genes based on KEGG metabolic pathways [Li and Li, 2008]: Y is time to death (Glioblastoma) and X is gene expression. A network is given on the p genes based on KEGG metabolic pathways Nathalie Vialaneix | Using network in prediction models 12/22
  • 18. Sketch of main directions 1 use a kernel based on the Laplacian and its associated dot product to compute a distance 2 use a standard linear model but regularize/penalize it with the Laplacian norm Nathalie Vialaneix | Using network in prediction models 13/22
  • 19. Working in the feature space... [Rapaport et al., 2007] xi is gene expression (length p) for sample i Standard approach → a gene expression profile: xi = p X j=1 xij                          0 . . . 1 . . . 0                          → dot product is: p X j=1 (xijxi0j) Nathalie Vialaneix | Using network in prediction models 14/22
  • 20. Working in the feature space... [Rapaport et al., 2007] xi is gene expression (length p) for sample i Standard approach → a gene expression profile: xi = p X j=1 xij                          0 . . . 1 . . . 0                          → dot product is: p X j=1 (xijxi0j) Laplacian approach → a gene expression profile: Φ(xi) = p X j=1 φ(λj)x̃ijej with x̃ij = e j xi (similar to Fourier transform) → dot product is: p X j=1 x i Kφxi0 with Kφ the kernel obtained from L with φ Nathalie Vialaneix | Using network in prediction models 14/22
  • 21. Working in the feature space... and then? [Rapaport et al., 2007] ∼ Standard linear regression with the Laplacian dot product + `2 penalty with the Laplacian kernel norm: arg min w∈Rp n X i=1 w Φ(xi) − yi 2 +Ckwk2 ⇔ arg min β∈Rp n X i=1 β xi − yi 2 +Cβ K−1 φ β this is called a SVM...! redundant information from two genes that are strongly connected on the graph is taken into account: this approach enforces similar contributions of these genes for the prediction improved interpretation Nathalie Vialaneix | Using network in prediction models 15/22
  • 22. Direct penalty with L [Li and Li, 2008] arg min β∈Rp n X i=1 β xi − yi 2 + Cβ Lβ +C0 kβk1 | {z } to enforce sparsity Nathalie Vialaneix | Using network in prediction models 16/22
  • 23. Direct penalty with L [Li and Li, 2008] arg min β∈Rp n X i=1 β xi − yi 2 + Cβ Lβ +C0 kβk1 | {z } to enforce sparsity Remarks: this is a very similar idea than previously (identical penalty for φ(·) = 1/·). Nathalie Vialaneix | Using network in prediction models 16/22
  • 24. Direct penalty with L [Li and Li, 2008] arg min β∈Rp n X i=1 β xi − yi 2 + Cβ Lβ +C0 kβk1 | {z } to enforce sparsity Remarks: this is a very similar idea than previously (identical penalty for φ(·) = 1/·). [Li and Li, 2010] explain that this forces connected vertices to have similar contributions whereas in some cases, they can have opposite contributions ⇒ use of L β jj0 =          −sign(βj)sign(βj0 ) if j , j0 and vj and vj0 are linked by an edge 0 if j , j0 and vj and vj0 are not linked by an edge dj if j = j0 Nathalie Vialaneix | Using network in prediction models 16/22
  • 25. Outline 1 A very short introduction 2 Some background on graphs and kernels 3 Using Laplacian in prediction 4 Coming back to our problem... what about SNP networks? Nathalie Vialaneix | Using network in prediction models 17/22
  • 26. Similar approach... with SNPs [Azencott et al., 2013] Background: X are SNP values, Y is a phenotype but it is not used as but to derive a association score between each single SNP and the phenotype, c ∈ Rp (ex: SKAT kernel) In addition, a network is given between SNPs (in the paper, derived from KEGG gene network or co-expression network) Nathalie Vialaneix | Using network in prediction models 18/22
  • 27. Similar approach... with SNPs [Azencott et al., 2013] Background: X are SNP values, Y is a phenotype but it is not used as but to derive a association score between each single SNP and the phenotype, c ∈ Rp (ex: SKAT kernel) In addition, a network is given between SNPs (in the paper, derived from KEGG gene network or co-expression network) Purpose: discover sets of genetic loci that are maximally associated with a phenotype while being connected in the underlying network → search for s ∈ {0, 1}p with the association score associated to s given by Pp j=1 sjcj = sc (to be maximized) Nathalie Vialaneix | Using network in prediction models 18/22
  • 28. Similar approach... with SNPs [Azencott et al., 2013] Background: X are SNP values, Y is a phenotype but it is not used as but to derive a association score between each single SNP and the phenotype, c ∈ Rp (ex: SKAT kernel) In addition, a network is given between SNPs (in the paper, derived from KEGG gene network or co-expression network) Purpose: discover sets of genetic loci that are maximally associated with a phenotype while being connected in the underlying network → search for s ∈ {0, 1}p with the association score associated to s given by Pp j=1 sjcj = sc (to be maximized) Solution: arg max s∈{0,1}p s c − Cs Ls −C0 ksk0 | {z } to enforce sparsity Nathalie Vialaneix | Using network in prediction models 18/22
  • 29. Another similar reference... with SNPs [Chiquet et al., 2016] Background: X are SNP values, Y are several phenotypes (multi-trait prediction ⇒ multivariate regression model) In addition, the authors do not directly use a network but a similarity that is related to the genomic distance, L Nathalie Vialaneix | Using network in prediction models 19/22
  • 30. Another similar reference... with SNPs [Chiquet et al., 2016] Background: X are SNP values, Y are several phenotypes (multi-trait prediction ⇒ multivariate regression model) In addition, the authors do not directly use a network but a similarity that is related to the genomic distance, L Purpose: prediction, estimation of coefficients (in linear models) and of direct effects are obtained using a model with i/ `2 regularization with the norm induced by L and 2/ `1 penalization to enforce sparsity (e.g., select SNPs) Nathalie Vialaneix | Using network in prediction models 19/22
  • 31. Other references on SNP networks [Hu et al., 2011] their aim is SNP network construction in association with a phenotype (for further mining): pairwise interactions detected with entropy gain (cancer) [Kogelman and Kadarmideen, 2014] their aim is also SNP network construction in association with a phenotype (for clustering + mining): combine genomic correlation (similar to LD) with genetic association (interaction tests) in a very crude way (almost just a sum) (pig carcass weight) Nathalie Vialaneix | Using network in prediction models 20/22
  • 32. Promising questions using a priori SNP network in association models comparing SNP networks (as described in the previous slide) with a priori networks to check if this solution is potentially interesting ... that will be for the next meeting! Nathalie Vialaneix | Using network in prediction models 21/22
  • 33. Nathalie Vialaneix | Using network in prediction models 22/22
  • 34. References Azencott, C.-A., Grimm, D., Sugiyama, M., Kawahara, Y., and Borgwardt, K. M. (2013). Efficient network-guided multi-locus association mapping with graph cuts. Bioinformatics, 29(13):i171–i179. Chiquet, J., Mary-Huard, T., and Robin, S. (2016). Structured regularization for conditional Gaussian graphical models. Statistics and Computing, pages 789–804. Fouss, F., Pirotte, A., Renders, J., and Saerens, M. (2007). Random-walk computation of similarities between nodes of a graph, with application to collaborative recommendation. IEEE Transactions on Knowledge and Data Engineering, 19(3):355–369. Hu, T., Sinnott-Armstrong, N. A., Kiralis, J. W., Andrew, A. S., Karagas, M. R., and Moore, J. H. (2011). Characterizing genetic interactions in human disease association studies using statistical epistasis networks. BMC Bioinformatics, 12:364. Kogelman, L. J. and Kadarmideen, H. N. (2014). Weighted interaction SNP hub (WISH) network method for building genetic networks for complex diseases traits using whole genome genotype data. BMC Systems Biology, 8(Suppl 2). Kondor, R. and Lafferty, J. (2002). Diffusion kernels on graphs and other discrete structures. In Sammut, C. and Hoffmann, A., editors, Proceedings of the 19th International Conference on Machine Learning, pages 315–322, Sydney, Australia. Morgan Kaufmann Publishers Inc. San Francisco, CA, USA. Li, C. and Li, H. (2008). Network-constrained regularization and variable selection for analysis of genomic data. Bioinformatics, 24(9):1175–1182. Li, C. and Li, H. (2010). Variable selection and regression analysis for graph-structured covariates with an application to genomics. The Annals of Applied Statistics, 4(3):1498–1516. Nathalie Vialaneix | Using network in prediction models 22/22
  • 35. Rapaport, F., Zinovyev, A., Dutreix, M., Barillot, E., and Vert, J. (2007). Classification of microarray data using gene networks. BMC Bioinformatics, 8:35. Smola, A. and Kondor, R. (2003). Kernels and regularization on graphs. In Warmuth, M. and Schölkopf, B., editors, Proceedings of the Conference on Learning Theory (COLT) and Kernel Workshop, Lecture Notes in Computer Science, pages 144–158, Washington, DC, USA. Springer-Verlag Berlin Heidelberg. von Luxburg, U. (2007). A tutorial on spectral clustering. Statistics and Computing, 17(4):395–416. Nathalie Vialaneix | Using network in prediction models 22/22