Direct Sampling Direct Sampling is an approximate inference method where we pull samples from the given joint probability distribution. Example Suppose we are interested in: where we dare P(B^{1}|D^{1},C^{1}). Step 1: sort We obtain a topological sort of this network:

\begin{equation} B, S, E, D, C \end{equation}

Step 2: sample from B,S We sample B. We sampled that B=1 today. We sample S. We sampled that S=0 today. Step 3: sample from E We sample E GIVEN what we already sampled, that B=1, S=0, we sampled that that E = 1 Step 4: sample from D, C We sample D given that E=1 as we sampled. We sample C given that E=1 as we sampled. Repeat Repeat steps 2-4 Step n: Analyze B S E D C 1 0 1 0 1 0 1 1 0 0 1 1 1 1 0 0 0 1 1 0 1 0 1 1 1 We desire to know P(b^{1}|d^{1}, c^{1}). Looks like, given this table, it would be 100\%. Likelihood Weighted Sampling Likelihood Weighted Sampling is a sampling approach whereby you force values that you wont, and then weight the results by the chance of it happening. This is super useful when our envidence is unlikely. Example Suppose again you are interested in P(b^{1}|d^{1}, c^{1}). In this case, we only sample B,S,E: B S E 0 1 0 1 0 1 Now, for each of these results, we the compute the chance of our priors happening given the samples. Row 1: p(d^{1}|e^{0})p(c^{1}|e^{0}) Row 2: p(d^{1}|e^{1})p(c^{1}|e^{1}) Let’s say: Row 1: p(d^{1}|e^{0})p(c^{1}|e^{0})=0.3 Row 2: p(d^{1}|e^{1})p(c^{1}|e^{1})=0.9 Finally, to compute p(b^{1}|d^{1}c^{1}):

\begin{equation} \frac{0.9}{0.9+0.3} \end{equation}

because only row 2 fit with our expectations.

[[curator]]
I'm the Curator. I can help you navigate, organize, and curate this wiki. What would you like to do?