SlideShare a Scribd company logo
Gaussian Process Regression
Random Process
A random process 𝑋𝑡 is completely characterized if the following is known.
𝑃((𝑋𝑡1
, ⋯ ⋯ , 𝑋𝑡 𝑘
) for any 𝐵, 𝑘, and 𝑡1, ⋯ ⋯ , 𝑡 𝑘
A random process (RP) (or stochastic process) is an infinite indexed collection
of random variables {𝑋(𝑡) ∶ 𝑡 ∈ 𝑇 }, defined over a common probability space.
(Functions are infinite dimensional vectors)
Note that given a random process, only ’finite-dimensional’ probabilities or
probability functions can be specified
𝐹𝑜𝑟 𝑡𝑖𝑚𝑒 𝑡 ∈ 𝑇 𝑎𝑛𝑑 𝑡ℎ𝑒 𝑜𝑢𝑡𝑐𝑜𝑚𝑒 𝑜𝑓 𝑡ℎ𝑒 𝑢𝑛𝑑𝑒𝑟𝑙𝑦𝑖𝑛𝑔 𝑟𝑎𝑛𝑑𝑜𝑚 𝑒𝑥𝑝𝑒𝑟𝑖𝑚𝑒𝑛𝑡 𝜔 ∈ Ω
𝑇 × Ω → ℝ
Random Process
Gaussian Process
(Background)
Gaussian Process
Gaussian Process
A Gaussian process is a collection of random variables, any finite number of
which have a joint Gaussian distribution
Gaussian Process
Gaussian Process
* Multivariate and Joint distribution are basically synonyms.
Gaussian Process
Gaussian Process
Gaussian process and Gaussian process regression are different.
Gaussian process regression: A nonparametric Bayesian
regression method using the properties of Gaussian processes.
Two views to interpret Gaussian process regression
• Weight-space view
• Function-space view
MLE vs MAP
Linear regression, 𝑓 𝑥 = 𝑤 𝑇 𝑥
𝐺𝑜𝑎𝑙 𝑜𝑓 𝑙𝑖𝑛𝑒𝑎𝑟 𝑟𝑒𝑔𝑟𝑒𝑠𝑠𝑖𝑜𝑛
𝑚𝑖𝑛𝑚𝑖𝑧𝑒: 𝑦 − 𝑓(𝑥) 2
𝑠𝑜𝑙𝑢𝑡𝑖𝑜𝑛: 𝑤 = (𝑋𝑋 𝑇
)−1
𝑋𝑦
MLE vs MAP
MLE vs MAP
Another perspective of Bayesian linear regression :
Ridge regularization
MLE vs MAP
MLE vs MAP
MLE vs MAP
Return to Bayesian solution:
Mean value of 𝑥𝑤 𝑀𝐴𝑃
Gaussian Process regression
Gaussian Process regression
• Weight Space View
• Function Space View
Weight Space View
Gaussian Process regression
Weight Space View
𝑌 = ∅(𝑥) 𝑇 𝑤, 𝑤~𝑁 0, 𝐴−1 𝐼
𝐼 𝑚𝑒𝑎𝑛𝑠 𝑡ℎ𝑒𝑟𝑒 𝑖𝑠 𝑛𝑜
𝑐𝑜𝑙𝑖𝑛𝑒𝑎𝑟𝑡𝑦 𝑖𝑛 𝑐𝑜𝑣𝑎𝑟𝑖𝑎𝑡𝑒𝑠 𝑜𝑓 𝑤
𝐸 𝑌 = 𝐸 ∅ 𝑥 𝑇 𝑤 = ∅ 𝑥 𝑇 𝐸 𝑤 = 0
Cov 𝑌 = 𝐸 𝑌 − 0 𝑌 − 0 = 𝐸 𝑌𝑌 𝑇
= ∅ 𝑥 𝑇 𝐸 𝑤𝑤 𝑇 ∅ 𝑥 = ∅ 𝑥 𝑇 𝐴−1∅ 𝑥
𝑘 𝑥𝑖, 𝑥𝑗 = 𝑒𝑥𝑝(− 𝑥𝑖 − 𝑥𝑗
2
)
𝑘 𝑋 𝑇, 𝑋 = ∅ 𝑥 𝑇∅ 𝑥 =
𝑘(𝑥1, 𝑥1) ⋯ 𝑘(𝑥1, 𝑥 𝑛)
⋮ ⋱ ⋮
𝑘(𝑥 𝑛, 𝑥1) ⋯ 𝑘(𝑥 𝑛, 𝑥 𝑛)
𝑤𝑒 𝑑𝑒𝑓𝑖𝑛𝑒 𝐾 = ∅ 𝑥 𝑇 𝐴−1∅ 𝑥
𝑃 𝑌 = 𝑁(𝑌ㅣ0, 𝐾)
𝑤ℎ𝑎𝑡 𝑑𝑜 𝑤𝑒 𝑔𝑒𝑡 𝑖𝑓 𝑛𝑒𝑤 𝑑𝑎𝑡𝑎 𝑥∗ 𝑎𝑝𝑝𝑒𝑎𝑟?
𝑃 𝑦∗ㅣ𝑥∗, 𝑋, 𝑌 = 𝑁(𝑦∗ㅣ? , ? )
Weight Space View
𝑤ℎ𝑎𝑡 𝑑𝑜 𝑤𝑒 𝑔𝑒𝑡 𝑖𝑓 𝑛𝑒𝑤 𝑑𝑎𝑡𝑎 𝑥∗ ℎ𝑎𝑠 𝑎𝑝𝑝𝑒𝑎𝑟?
𝑃 𝑦∗ㅣ𝑥∗, 𝑋, 𝑌 = 𝑁(𝑦∗ㅣ? , ? )
𝑊ℎ𝑎𝑡 𝑎𝑟𝑒 𝑡ℎ𝑒 𝑟𝑎𝑛𝑑𝑜𝑚 𝑣𝑎𝑟𝑖𝑎𝑏𝑙𝑒𝑠?
𝑃 𝑦∗ㅣ𝑌
Weight Space View
∑ 𝑦∗,𝑌 = 𝐸 𝑦∗ 𝑌 𝑇 = ∅(𝑥∗) 𝑇 𝐴−1∅(𝑋) 𝑇 = 𝐶
∑ 𝑦∗,𝑦∗ = 𝐸 𝑦∗ 𝑦∗ 𝑇
= ∅(𝑥∗) 𝑇 𝐴−1∅(𝑥∗) 𝑇 =k
𝑃 𝑦∗ㅣ𝑌 = 𝑃 ∅(𝑥∗) 𝑇 𝑤ㅣ∅(𝑋) 𝑇 𝑤 = 𝑁(𝑦∗ㅣ𝜇 𝑦∗ + ∑ 𝑦∗,𝑌 ∑ 𝑌,𝑌
−1
𝑌 − 𝜇 𝑌 , ∑ 𝑦∗,𝑦∗ −∑ 𝑦∗,𝑌 ∑ 𝑌,𝑌
−1
∑ 𝑌,𝑦∗)
= 𝑁(𝑦∗ㅣ𝐶𝐾−1 𝑌, 𝑘 − 𝐶𝐾−1 𝐶 𝑇)
Var( 𝑌∗)?
𝑌∗ = [𝑦1, ⋯ , 𝑦𝑛, 𝑦 𝑛+1] 𝑇, → Var( 𝑌∗)= 𝑐𝑜𝑣 𝑛 𝐶 𝑇
𝐶 𝑘
Weight Space View
𝑊ℎ𝑎𝑡 𝑖𝑓 𝑌 = ∅(𝑥) 𝑇
𝑤 + 𝜀, 𝑤~𝑁 0, 𝐴−1
𝐼 , 𝜀~𝑁 0, 𝐵−1
𝐼
Cov 𝑌 = 𝐸 𝑌 − 0 𝑌 − 0 = 𝐸 𝑌𝑌 𝑇 = ∅ 𝑥 𝑇 𝐸 𝑤𝑤 𝑇 ∅ 𝑥 + 𝐸 2𝜀∅ 𝑥 𝑇 𝑤 + 𝜀𝜀 𝑇
= ∅ 𝑥 𝑇 𝐴−1∅ 𝑥 + 𝐵−1 𝐼 = K + 𝐵−1 𝐼
Function Space View
Gaussian Process regression
Function Space View
Function Space View
Function Space View
Function Space View
Function Space View
Function Space View
Function Space View
Function Space View
Amazing properties of Non-parametric method
References
References
References
C. E. Rasmussen and C. K. Williams. Gaussian processes for machine learning, volume 1.
MIT press Cambridge, 2006.
References
"Gaussian Process", Lectured by Professor Il-Chul Moon
-video link: https://youtu.be/RmN54ykspK4
Ian Goodfellow et al. Deep Learning, (2016)
Trevor Hastie et al. The Elements of Statistical Learning (2001)
Machine Learning Lecture 26 "Gaussian Processes" -Cornell CS4780 SP17 by Kilian Weinberger
-video link: https://www.youtube.com/watch?v=R-NUdqxKjos&t=1000s
9.520/6.860S Statistical Learning Theory by Lorenzo Rosasco
http://www.mit.edu/~9.520/fall14/slides/class03/class03_rkhsPart1.pdf
-video link: https://www.youtube.com/watch?v=9-oxo_k69qs
Bayesian Deep Learning by Sungjoon Choi
-video link: https://www.edwith.org/bayesiandeeplearning/joinLectures/14426
Gaussian Process Regression

More Related Content

PDF
A brief introduction to Gaussian process
PDF
Gaussian Processes: Applications in Machine Learning
PPT
Back propagation
PPTX
The n Queen Problem
PPT
Branch and bound
PDF
Lec3 dqn
PDF
Markov Chain Monte Carlo Methods
PPTX
Perceptron & Neural Networks
A brief introduction to Gaussian process
Gaussian Processes: Applications in Machine Learning
Back propagation
The n Queen Problem
Branch and bound
Lec3 dqn
Markov Chain Monte Carlo Methods
Perceptron & Neural Networks

What's hot (20)

PDF
02 Machine Learning - Introduction probability
PPTX
Introduction to Machine Learning
PPTX
K Nearest Neighbor Algorithm
PDF
Optimization for Neural Network Training - Veronica Vilaplana - UPC Barcelona...
PPTX
Bayesian Neural Networks
PPTX
Random Forest
PPTX
UNIT III (8).pptx
PDF
AI_Unit I notes .pdf
PDF
Introduction to Recurrent Neural Network
PDF
Recurrent Neural Networks
PPTX
Lecture 18: Gaussian Mixture Models and Expectation Maximization
PPTX
data structures- back tracking
PPTX
Chapter 5 of 1
PPT
First order logic in knowledge representation
PDF
First Order Logic resolution
PDF
Dimensionality Reduction
PPTX
Reinforcement Learning
PPTX
Firefly algorithm
PPTX
AlexNet(ImageNet Classification with Deep Convolutional Neural Networks)
PPTX
An introduction to reinforcement learning
02 Machine Learning - Introduction probability
Introduction to Machine Learning
K Nearest Neighbor Algorithm
Optimization for Neural Network Training - Veronica Vilaplana - UPC Barcelona...
Bayesian Neural Networks
Random Forest
UNIT III (8).pptx
AI_Unit I notes .pdf
Introduction to Recurrent Neural Network
Recurrent Neural Networks
Lecture 18: Gaussian Mixture Models and Expectation Maximization
data structures- back tracking
Chapter 5 of 1
First order logic in knowledge representation
First Order Logic resolution
Dimensionality Reduction
Reinforcement Learning
Firefly algorithm
AlexNet(ImageNet Classification with Deep Convolutional Neural Networks)
An introduction to reinforcement learning
Ad

Similar to Gaussian Process Regression (20)

PDF
Differential Geometry for Machine Learning
PDF
Strong convexity on gradient descent and newton's method
PPTX
Solution of equations and eigenvalue problems
PPTX
Solving Poisson Equation using Conjugate Gradient Method and its implementation
PDF
Dual Spaces of Generalized Cesaro Sequence Space and Related Matrix Mapping
PDF
AJMS_482_23.pdf
PPTX
Conformal Boundary conditions
PPTX
Learning group em - 20171025 - copy
PDF
Stochastic optimal control & rl
PDF
QTML2021 UAP Quantum Feature Map
PPTX
GATE Engineering Maths : Vector Calculus
PPTX
tut07.pptx
PDF
Periodic Solutions for Nonlinear Systems of Integro-Differential Equations of...
PPTX
20180831 riemannian representation learning
PDF
Basic calculus (ii) recap
PPTX
Variational Autoencoder Tutorial
PDF
A Non Local Boundary Value Problem with Integral Boundary Condition
Differential Geometry for Machine Learning
Strong convexity on gradient descent and newton's method
Solution of equations and eigenvalue problems
Solving Poisson Equation using Conjugate Gradient Method and its implementation
Dual Spaces of Generalized Cesaro Sequence Space and Related Matrix Mapping
AJMS_482_23.pdf
Conformal Boundary conditions
Learning group em - 20171025 - copy
Stochastic optimal control & rl
QTML2021 UAP Quantum Feature Map
GATE Engineering Maths : Vector Calculus
tut07.pptx
Periodic Solutions for Nonlinear Systems of Integro-Differential Equations of...
20180831 riemannian representation learning
Basic calculus (ii) recap
Variational Autoencoder Tutorial
A Non Local Boundary Value Problem with Integral Boundary Condition
Ad

More from SEMINARGROOT (20)

PDF
Metric based meta_learning
PDF
Sampling method : MCMC
PDF
Demystifying Neural Style Transfer
PDF
Towards Deep Learning Models Resistant to Adversarial Attacks.
PDF
The ways of node embedding
PDF
Graph Convolutional Network
PDF
Denoising With Frequency Domain
PDF
Bayesian Statistics
PDF
Coding Test Review 3
PDF
Time Series Analysis - ARMA
PDF
Generative models : VAE and GAN
PDF
Effective Python
PDF
Understanding Blackbox Prediction via Influence Functions
PDF
Attention Is All You Need
PDF
Attention
PDF
WWW 2020 XAI Tutorial Review
PDF
Coding test review 2
PDF
Locality sensitive hashing
PDF
Coding Test Review1
PDF
SVM (Support Vector Machine & Kernel)
Metric based meta_learning
Sampling method : MCMC
Demystifying Neural Style Transfer
Towards Deep Learning Models Resistant to Adversarial Attacks.
The ways of node embedding
Graph Convolutional Network
Denoising With Frequency Domain
Bayesian Statistics
Coding Test Review 3
Time Series Analysis - ARMA
Generative models : VAE and GAN
Effective Python
Understanding Blackbox Prediction via Influence Functions
Attention Is All You Need
Attention
WWW 2020 XAI Tutorial Review
Coding test review 2
Locality sensitive hashing
Coding Test Review1
SVM (Support Vector Machine & Kernel)

Recently uploaded (20)

PPTX
UNIT 4 Total Quality Management .pptx
PPTX
Foundation to blockchain - A guide to Blockchain Tech
PPT
Project quality management in manufacturing
PPTX
Internet of Things (IOT) - A guide to understanding
PDF
R24 SURVEYING LAB MANUAL for civil enggi
PPTX
KTU 2019 -S7-MCN 401 MODULE 2-VINAY.pptx
PPTX
M Tech Sem 1 Civil Engineering Environmental Sciences.pptx
PDF
Enhancing Cyber Defense Against Zero-Day Attacks using Ensemble Neural Networks
PDF
TFEC-4-2020-Design-Guide-for-Timber-Roof-Trusses.pdf
PDF
Operating System & Kernel Study Guide-1 - converted.pdf
PPTX
Sustainable Sites - Green Building Construction
PDF
Model Code of Practice - Construction Work - 21102022 .pdf
DOCX
ASol_English-Language-Literature-Set-1-27-02-2023-converted.docx
PDF
SM_6th-Sem__Cse_Internet-of-Things.pdf IOT
PPTX
IOT PPTs Week 10 Lecture Material.pptx of NPTEL Smart Cities contd
PPTX
Infosys Presentation by1.Riyan Bagwan 2.Samadhan Naiknavare 3.Gaurav Shinde 4...
PPTX
Engineering Ethics, Safety and Environment [Autosaved] (1).pptx
DOCX
573137875-Attendance-Management-System-original
PPTX
CARTOGRAPHY AND GEOINFORMATION VISUALIZATION chapter1 NPTE (2).pptx
PPTX
MET 305 2019 SCHEME MODULE 2 COMPLETE.pptx
UNIT 4 Total Quality Management .pptx
Foundation to blockchain - A guide to Blockchain Tech
Project quality management in manufacturing
Internet of Things (IOT) - A guide to understanding
R24 SURVEYING LAB MANUAL for civil enggi
KTU 2019 -S7-MCN 401 MODULE 2-VINAY.pptx
M Tech Sem 1 Civil Engineering Environmental Sciences.pptx
Enhancing Cyber Defense Against Zero-Day Attacks using Ensemble Neural Networks
TFEC-4-2020-Design-Guide-for-Timber-Roof-Trusses.pdf
Operating System & Kernel Study Guide-1 - converted.pdf
Sustainable Sites - Green Building Construction
Model Code of Practice - Construction Work - 21102022 .pdf
ASol_English-Language-Literature-Set-1-27-02-2023-converted.docx
SM_6th-Sem__Cse_Internet-of-Things.pdf IOT
IOT PPTs Week 10 Lecture Material.pptx of NPTEL Smart Cities contd
Infosys Presentation by1.Riyan Bagwan 2.Samadhan Naiknavare 3.Gaurav Shinde 4...
Engineering Ethics, Safety and Environment [Autosaved] (1).pptx
573137875-Attendance-Management-System-original
CARTOGRAPHY AND GEOINFORMATION VISUALIZATION chapter1 NPTE (2).pptx
MET 305 2019 SCHEME MODULE 2 COMPLETE.pptx

Gaussian Process Regression

  • 2. Random Process A random process 𝑋𝑡 is completely characterized if the following is known. 𝑃((𝑋𝑡1 , ⋯ ⋯ , 𝑋𝑡 𝑘 ) for any 𝐵, 𝑘, and 𝑡1, ⋯ ⋯ , 𝑡 𝑘 A random process (RP) (or stochastic process) is an infinite indexed collection of random variables {𝑋(𝑡) ∶ 𝑡 ∈ 𝑇 }, defined over a common probability space. (Functions are infinite dimensional vectors) Note that given a random process, only ’finite-dimensional’ probabilities or probability functions can be specified 𝐹𝑜𝑟 𝑡𝑖𝑚𝑒 𝑡 ∈ 𝑇 𝑎𝑛𝑑 𝑡ℎ𝑒 𝑜𝑢𝑡𝑐𝑜𝑚𝑒 𝑜𝑓 𝑡ℎ𝑒 𝑢𝑛𝑑𝑒𝑟𝑙𝑦𝑖𝑛𝑔 𝑟𝑎𝑛𝑑𝑜𝑚 𝑒𝑥𝑝𝑒𝑟𝑖𝑚𝑒𝑛𝑡 𝜔 ∈ Ω 𝑇 × Ω → ℝ
  • 5. Gaussian Process A Gaussian process is a collection of random variables, any finite number of which have a joint Gaussian distribution
  • 7. Gaussian Process * Multivariate and Joint distribution are basically synonyms.
  • 9. Gaussian Process Gaussian process and Gaussian process regression are different. Gaussian process regression: A nonparametric Bayesian regression method using the properties of Gaussian processes. Two views to interpret Gaussian process regression • Weight-space view • Function-space view
  • 10. MLE vs MAP Linear regression, 𝑓 𝑥 = 𝑤 𝑇 𝑥 𝐺𝑜𝑎𝑙 𝑜𝑓 𝑙𝑖𝑛𝑒𝑎𝑟 𝑟𝑒𝑔𝑟𝑒𝑠𝑠𝑖𝑜𝑛 𝑚𝑖𝑛𝑚𝑖𝑧𝑒: 𝑦 − 𝑓(𝑥) 2 𝑠𝑜𝑙𝑢𝑡𝑖𝑜𝑛: 𝑤 = (𝑋𝑋 𝑇 )−1 𝑋𝑦
  • 12. MLE vs MAP Another perspective of Bayesian linear regression : Ridge regularization
  • 15. MLE vs MAP Return to Bayesian solution: Mean value of 𝑥𝑤 𝑀𝐴𝑃
  • 16. Gaussian Process regression Gaussian Process regression • Weight Space View • Function Space View
  • 17. Weight Space View Gaussian Process regression
  • 18. Weight Space View 𝑌 = ∅(𝑥) 𝑇 𝑤, 𝑤~𝑁 0, 𝐴−1 𝐼 𝐼 𝑚𝑒𝑎𝑛𝑠 𝑡ℎ𝑒𝑟𝑒 𝑖𝑠 𝑛𝑜 𝑐𝑜𝑙𝑖𝑛𝑒𝑎𝑟𝑡𝑦 𝑖𝑛 𝑐𝑜𝑣𝑎𝑟𝑖𝑎𝑡𝑒𝑠 𝑜𝑓 𝑤 𝐸 𝑌 = 𝐸 ∅ 𝑥 𝑇 𝑤 = ∅ 𝑥 𝑇 𝐸 𝑤 = 0 Cov 𝑌 = 𝐸 𝑌 − 0 𝑌 − 0 = 𝐸 𝑌𝑌 𝑇 = ∅ 𝑥 𝑇 𝐸 𝑤𝑤 𝑇 ∅ 𝑥 = ∅ 𝑥 𝑇 𝐴−1∅ 𝑥 𝑘 𝑥𝑖, 𝑥𝑗 = 𝑒𝑥𝑝(− 𝑥𝑖 − 𝑥𝑗 2 ) 𝑘 𝑋 𝑇, 𝑋 = ∅ 𝑥 𝑇∅ 𝑥 = 𝑘(𝑥1, 𝑥1) ⋯ 𝑘(𝑥1, 𝑥 𝑛) ⋮ ⋱ ⋮ 𝑘(𝑥 𝑛, 𝑥1) ⋯ 𝑘(𝑥 𝑛, 𝑥 𝑛) 𝑤𝑒 𝑑𝑒𝑓𝑖𝑛𝑒 𝐾 = ∅ 𝑥 𝑇 𝐴−1∅ 𝑥 𝑃 𝑌 = 𝑁(𝑌ㅣ0, 𝐾) 𝑤ℎ𝑎𝑡 𝑑𝑜 𝑤𝑒 𝑔𝑒𝑡 𝑖𝑓 𝑛𝑒𝑤 𝑑𝑎𝑡𝑎 𝑥∗ 𝑎𝑝𝑝𝑒𝑎𝑟? 𝑃 𝑦∗ㅣ𝑥∗, 𝑋, 𝑌 = 𝑁(𝑦∗ㅣ? , ? )
  • 19. Weight Space View 𝑤ℎ𝑎𝑡 𝑑𝑜 𝑤𝑒 𝑔𝑒𝑡 𝑖𝑓 𝑛𝑒𝑤 𝑑𝑎𝑡𝑎 𝑥∗ ℎ𝑎𝑠 𝑎𝑝𝑝𝑒𝑎𝑟? 𝑃 𝑦∗ㅣ𝑥∗, 𝑋, 𝑌 = 𝑁(𝑦∗ㅣ? , ? ) 𝑊ℎ𝑎𝑡 𝑎𝑟𝑒 𝑡ℎ𝑒 𝑟𝑎𝑛𝑑𝑜𝑚 𝑣𝑎𝑟𝑖𝑎𝑏𝑙𝑒𝑠? 𝑃 𝑦∗ㅣ𝑌
  • 20. Weight Space View ∑ 𝑦∗,𝑌 = 𝐸 𝑦∗ 𝑌 𝑇 = ∅(𝑥∗) 𝑇 𝐴−1∅(𝑋) 𝑇 = 𝐶 ∑ 𝑦∗,𝑦∗ = 𝐸 𝑦∗ 𝑦∗ 𝑇 = ∅(𝑥∗) 𝑇 𝐴−1∅(𝑥∗) 𝑇 =k 𝑃 𝑦∗ㅣ𝑌 = 𝑃 ∅(𝑥∗) 𝑇 𝑤ㅣ∅(𝑋) 𝑇 𝑤 = 𝑁(𝑦∗ㅣ𝜇 𝑦∗ + ∑ 𝑦∗,𝑌 ∑ 𝑌,𝑌 −1 𝑌 − 𝜇 𝑌 , ∑ 𝑦∗,𝑦∗ −∑ 𝑦∗,𝑌 ∑ 𝑌,𝑌 −1 ∑ 𝑌,𝑦∗) = 𝑁(𝑦∗ㅣ𝐶𝐾−1 𝑌, 𝑘 − 𝐶𝐾−1 𝐶 𝑇) Var( 𝑌∗)? 𝑌∗ = [𝑦1, ⋯ , 𝑦𝑛, 𝑦 𝑛+1] 𝑇, → Var( 𝑌∗)= 𝑐𝑜𝑣 𝑛 𝐶 𝑇 𝐶 𝑘
  • 21. Weight Space View 𝑊ℎ𝑎𝑡 𝑖𝑓 𝑌 = ∅(𝑥) 𝑇 𝑤 + 𝜀, 𝑤~𝑁 0, 𝐴−1 𝐼 , 𝜀~𝑁 0, 𝐵−1 𝐼 Cov 𝑌 = 𝐸 𝑌 − 0 𝑌 − 0 = 𝐸 𝑌𝑌 𝑇 = ∅ 𝑥 𝑇 𝐸 𝑤𝑤 𝑇 ∅ 𝑥 + 𝐸 2𝜀∅ 𝑥 𝑇 𝑤 + 𝜀𝜀 𝑇 = ∅ 𝑥 𝑇 𝐴−1∅ 𝑥 + 𝐵−1 𝐼 = K + 𝐵−1 𝐼
  • 22. Function Space View Gaussian Process regression
  • 31. Amazing properties of Non-parametric method
  • 33. References C. E. Rasmussen and C. K. Williams. Gaussian processes for machine learning, volume 1. MIT press Cambridge, 2006.
  • 34. References "Gaussian Process", Lectured by Professor Il-Chul Moon -video link: https://youtu.be/RmN54ykspK4 Ian Goodfellow et al. Deep Learning, (2016) Trevor Hastie et al. The Elements of Statistical Learning (2001) Machine Learning Lecture 26 "Gaussian Processes" -Cornell CS4780 SP17 by Kilian Weinberger -video link: https://www.youtube.com/watch?v=R-NUdqxKjos&t=1000s 9.520/6.860S Statistical Learning Theory by Lorenzo Rosasco http://www.mit.edu/~9.520/fall14/slides/class03/class03_rkhsPart1.pdf -video link: https://www.youtube.com/watch?v=9-oxo_k69qs Bayesian Deep Learning by Sungjoon Choi -video link: https://www.edwith.org/bayesiandeeplearning/joinLectures/14426