Bayesian Networks in Machine Learning

Overview of Other ML
Techniques
Geoff Hulten
On Being Bayesian
Conditional Independence
Bayesian Network
X
Y
Z
Lightning
Thunder
Rain
Represent conditional dependencies via a directed acyclic graph, where:
A variable is independent of its non-descendants given the value of its parents.
 
Eight Parameters
 
Five Parameters
 
Decompose joint distribution according to structure
 
3 binary variables
Inference in Bayesian Networks
Lightning
Rain
 
< Rain=0, Lightning=? >
 
< Rain=?, Lightning=1 >
 
= .1
 
Super simple case
 
Sorta simple case
 
Naïve Bayes
 
In general use techniques like EM or Gibbs sampling
More Complex Inference Situation
Lightning
Thunder
Rain
ForestFire
Campfire
TourGroup
Learning Bayesian Networks
 
Structure
Known
 
Structure
Unknown
 
All Variables Observed
 
Some Variables Hidden
 
MAP Estimates for Parameters
(Like Naïve Bayes)
 
Search for structure:
Initial State: empty network or prior network
Operations: Add arc, delete arc, reverse arc
Evaluations: LL(D|G) * prior(G)
 
EM Algorithm
 
Abandon Hope
 
Just a Joke:
Combine EM with
structure search…
 
Normalization
Example of Collaborative Filtering
 
Challenges:
Cold Start
Sparsity
Support Vector Machine (SVM)
Support Vector Machines
 
For non-linear data
Support Vector Machines (More Concepts)
 
Optimization
Solve constrained system of equations
Quadratic programming (e.g. SMO)
 
 
Dealing with noise (soft vs hard)
Summary
There are a lot of Machine Learning Algorithms…
Slide Note
Embed
Share

Bayesian Networks are probabilistic graphical models that represent relationships between variables. They are used for modeling uncertain knowledge and performing inference. This content covers topics such as conditional independence, representation of dependencies, inference techniques, and learning Bayesian networks using techniques like EM algorithm and structure search.

  • Bayesian Networks
  • Machine Learning
  • Probabilistic Models
  • Inference Techniques
  • EM Algorithm

Uploaded on Oct 04, 2024 | 0 Views


Download Presentation

Please find below an Image/Link to download the presentation.

The content on the website is provided AS IS for your information and personal use only. It may not be sold, licensed, or shared on other websites without obtaining consent from the author. Download presentation by click this link. If you encounter any issues during the download, it is possible that the publisher has removed the file from their server.

E N D

Presentation Transcript


  1. Overview of Other ML Techniques Geoff Hulten

  2. On Being Bayesian

  3. Conditional Independence Definition: X is conditionally independent of Y give Z if the probability distribution governing X is independent of the value of Y given the value of Z. ??,??,??? ? = ??? = ??,? = ?? = ? ? = ??? = ??) ? ? ?,? = ?(?|?) Example: Thunder is conditionally independent of Rain, given Lightning: ? ? ????? ????,??? ?????) = ? ? ????? ??? ?????)

  4. Bayesian Network Represent conditional dependencies via a directed acyclic graph, where: A variable is independent of its non-descendants given the value of its parents. 3 binary variables ?1 P(X=0,Y=0,Z=0) ?2 P(X=1,Y=0,Z=0) ?(? ?????,??? ?????,????) ? ?,?,? = Thunder is conditionally independent of Rain, given Lightning ?8 P(X=1,Y=1,Z=1) Z Eight Parameters X Y Five Parameters Rain Thunder ?1 ?1 P(Z=1) P(X=1|Z=0) ? ? ?,? = ?(?|?) ?2 P(X=1|Z=1) ? ? ?,? = ?(?|?) Lightning ?1 P(Y=1|Z=0) ?2 P(Y=1|Z=1) ? ?,?,? = ? ? ? ? ? ?(?|?) Decompose joint distribution according to structure

  5. Inference in Bayesian Networks ? = argmax ?(??) ?(??|??) ?? ? ? P(Rain) .3 Rain ? P(Lightning|Rain=0) .1 Lightning ?1 ? ?? P(Lightning|Rain=1) .5 Na ve Bayes Super simple case ~ P(Lightning|Rain=0) = .1 < Rain=0, Lightning=? > ?(???? = 1)?(??? ????? = 1|???? = 1) ? ?????(???? = ?)?(??? ????? = 1|???? = ?) ? ???? = 1 = ~ P(Rain |Lightning=1) < Rain=?, Lightning=1 > Sorta simple case In general use techniques like EM or Gibbs sampling

  6. More Complex Inference Situation def SimpleGibbs(???? ?????????,X,F): ???????= ? with unknown randomly initialized for i in range(burnInSteps): ???????= ??????(???????) samples = [] for i in range(len(samplesToGenerate)): ???????= ??????(???????) if i % samplesToSkip == 0: samples.append(???????) Rain TourGroup Lightning Campfire ForestFire Thunder return F(samples) ? ???? ?????????? = 1,? ????? = 0) = ?? def ??????(?): ????= ? for i in variables: if IsUnknown(??) ???? return ???? ? = ?????? ?????? ????? ????? ????

  7. Learning Bayesian Networks Just a Joke: Combine EM with structure search Search for structure: Unknown Structure Abandon Hope Initial State: empty network or prior network Operations: Add arc, delete arc, reverse arc Evaluations: LL(D|G) * prior(G) Structure Known MAP Estimates for Parameters (Like Na ve Bayes) EM Algorithm All Variables Observed Some Variables Hidden

  8. Normalization

  9. Example of Collaborative Filtering Will Alice like ?????2? ??????,??? = 1.38 ? = .357 ??????,? ??? = 1.41 ??????,2= 3.75 + .357(1.38 5 3.4 1.41 2 2.5 ) ??????,2= 3.75 + .357(2.22 + 0.70) Challenges: Cold Start Sparsity ??????,2= 3.75 + 1.04 ??????,2= 4.79

  10. Support Vector Machine (SVM) Linear separator ? ?^= ????(?0+ ?? ??) ?=1 Maximum Margin Support Vectors

  11. Support Vector Machines For non-linear data Data not linearly separable Add a new dimension: ????= 1.0 2(.5 ?1)2+(.5 ?2)2 Will be more linearly separable Various standard options (kernels) Try multiple methods Project back to original space

  12. Support Vector Machines (More Concepts) Optimization Solve constrained system of equations Quadratic programming (e.g. SMO) Dealing with noise (soft vs hard)

  13. Summary There are a lot of Machine Learning Algorithms

More Related Content

giItT1WQy@!-/#giItT1WQy@!-/#giItT1WQy@!-/#giItT1WQy@!-/#giItT1WQy@!-/#