Probabilistic Inference in Polytrees

Estudies4you
Probabilistic Inference in Polytrees of Artificial Intelligence

PROBABILITIC INFERENCE IN POLYTREES 

  • A polytree is a network in which there exists a node B such that some of the nodes are connected to it only through its parents. 
  • The nodes which are not connected by its parents are connected by its immediate successors. 
  • The nodes connected by parents are said to be above B. 
  • The nodes connected by its children are said to be below B. 
  • It has to be noted that in the polytree this B node is the only path between the above nodes and below nodes. 
  • The polytree network is shown in the Fig. 3.8.1.

Probabilistic Inference in Polytrees AI,examples of polytrees in AI,evidence of polytrees in AI,use of polytree in AI,what is polytree in AI jntuh,
Fig. 3.8.1 A Typical Polytree
Every node of the polytree has the same rules. 


Evidence Above 

    If all evidence nodes are above Q. We calculate P(B | A5, A4) as an example. . Here, bottom up recursive algorithm is used.
It is used to know the probability of every ancestor of B, if the evidence is given and till we reach the evidence or the evidence is below the ancestor.

The algorithm is as follows,. 
1) The parents of node B are taken, 
        P(B|{A_5},{A_5}) = \sum\limits_{{A_6},{A_7}} {P(B,{A_6},{A_7}{\rm{ |}}{\rm{ }}{{\rm{A}}_5},{A_4})}
        (Here the summation is used to mention the four combinations, one is what is present i e.,  A6, A7, second one is substituting ᆨA6 for A6, the other is substituting ᆨA7 for A7, and other is substituting ᆨA6. and- ᆨA7). 

2) Conditional independence is used make the parents of Q as part of the evidence. 
        P(B,{A_6},{A_7}{\rm{ |}}{\rm{ }}{A_5},{A_4}) = {\rm{ }}P(B,{A_6},{A_7}{\rm{,}}{A_5},{A_4}){\rm{ }}P({A_6},{A_7}{\rm{ | }}{A_5},{A_4})

3) Substituting this in first step gives, 
        P(B|{A_5},{A_4}) = \sum\limits_{{A_6},{A_7}} {P(B,{A_6},{A_7}{\rm{,}}{A_5},{A_4}){\rm{ P(}}{A_6},{A_7}{\rm{ | }}{{\rm{A}}_5},{A_4})}

4) As the node is conditionally independent of non descendants if it parents are given, 
        P(B{\rm{ |}}{\rm{ }}{A_5},{A_4}) = {\rm{ }}\sum\limits_{A6,A7} {P(B|{A_6},{A_7}){\rm{ }}P({A_6},{A_7}{\rm{ | }}{A_5},{A_6})}

5) By using D-separation we can split the parents, 
        P(B{\rm{ |}}{\rm{ }}{A_5},{A_4}) = {\rm{ }}\sum\limits_{A6,A7} {P(B|{A_6},{A_7}){\rm{ }}P({A_6}|{A_5}{\rm{,}}{A_6}){\rm{ P(}}{A_7}|{A_5}|{A_4})}

6) As it is allowed by D-separation we ignore the evidence which is above one parent while we are calculating the probability of another parent, 
        P(Q{\rm{ |}}{\rm{ }}{A_5},{A_4}) = {\rm{ }}\sum\limits_{A6,A7} {P(B|{A_6},{A_7}){\rm{ }}P({A_6}|{A_5}){\rm{ P(}}{A_7}|{A_4})}

In the above equation P(B | A6, A7) is the probability, of the query node, where the value of its parents are given. 
P(A6 | A5) and P(A7 | A4) are the probabilities of the parents if the part of the evidence above the parent is given. 
The recursive process of algorithm continues till we reach the node which has the evidence node as a parent. 
In P(A7 | A4) the A4 node is the evidence node and it is the parent of A7.
Now the parents are involved. 

Since, 
        P(A7,  A3 | A4) = P(A7 | A3, A4) P(A3 | A4) 

Then, 
        P({A_7}|{A_4}) = {\rm{ }}\sum\limits_{A3} {P({A_7}|{A_3},{A_4}){\rm{ }}P({A_3},{A_4})}

        P({A_7}|{A_4}) = {\rm{ }}\sum\limits_{A3} {P({A_7}|{A_3},{A_4}){\rm{ }}P({A_3})}

In P(A1 | A5) calculation we have, 
        P({A_6}|{A_5}) = {\rm{ }}\sum\limits_{A1,A2} {P({A_6}|{A_1},{A_2}){\rm{ }}P({A_1}|{A_5})} P({A_2})

As in P(A1 | A5) the evidence node is note above the query node this recursive procedure is terminated and. as it is below evidence below method has to be used.
Bayes rule can also be used' to Obtain P(Al | A5) 
        P({A_1}|{A_5}) = {\rm{ }}\frac{{P({A_5}|{A_1})P({A_1})}}{{P({A_5})}}

Now we can calculate P(A6 | A5) and then P(Q | A5, A4). 

Evidence Below

To calculate P(B | Al2, A13, A14, A11), as all the evidence nodes are below evidence below is used. 
Reasoning with Uncertain Information 
The algorithm used here is also a recursive algorithm. 
In this algorithm first we use Bayes rule. 
        {\rm{P(B | A12, A13, A14, A11)  =  }}\frac{{{\rm{P(A12, A13, A14, A11 | B) P(B)}}}}{{{\rm{P(A12, A13, A14, A11}})}}
Let, 
        K = {\rm{ }}\frac{1}{{P(A12,{\rm{ A13, A14, A11)}}}}

Now D-Separation is used. 
I({Al2, A13}, {A14, A11) | Q)

which gives, 
        P(B | Al2, A13, A14, A11) = KP(Al2, A13 | B) P(A14, A11 | B) P(B) 

Now as there is a single evidence node B we use top-down recursive algorithm instead of bottom-up approach. 

As A9 is the present of Al2, A13 
P(Al2, A13, A9 | B) = P(Al2, A13 | A9, Q) P(A9 | B) 

which is from the definition of conditional independence. So the equation becomes, 
        {\rm{P(A12, A13 | B)  =  }}\sum\limits_{A9} {{\rm{P(A12, A13 | A9, Q) P(A9 | B)}}}

By using D-Separation, 
        I{Al2, A13}, B | A9) 

SO, 
        {\rm{P(A12, A13 | B)  =  }}\sum\limits_{A9} {{\rm{P(A12, A13 | A9, Q) P(A9 | B)}}}

P(A9 | B) can calculated by the parents involved in it, 
        {\rm{P(A9 | B)  =  }}\sum\limits_{A8} {{\rm{P(A9 | A8, B) P(A8)}}}
The value of P(A9 | A8, B) is given by the network.

Now P(Al2, A13 | A9) has to be calculated in the same recursive fashion from top to down. 

The recursive call stops after one step as the children of A9 i.e. Al2 and A13 are evidence nodes because, and Al2 and A13 are independent if A9 is given. 
Now, 
        P(A12, A13 | A9) = P(A12 | A9) P(A13 | A9) 

Now we calculate P(A14, Al1 |  B) by using top-down, procedure, 
        {\rm{P(A14, A11 | B)  =  }}\sum\limits_{A10} {{\rm{P(A14, A11 | A10) P(A10 | B)}}}
        
        {\rm{P(A14, A11 | B)  =  }}\sum\limits_{A10} {{\rm{P(A14, A10), P(A11, A10) P(A10 | B)}}}
                                                                                                                [∴ I(A14, All A10)] 
In the above expression the value of the middle term is not given directly.

So, we calculate P(A11 | A10) by top-down approach, 
        {\rm{P(A11, A10)  =  }}\sum\limits_{A15} {{\rm{P(A11 | A15, A10), P(A15 | A10)}}}

        {\rm{P(A15, A10)  =  }}\sum\limits_{A11} {{\rm{P(A15 | A10, A11), P(A11)}}}
In P(A11 | A15, A10). All is above the evidence nodes, so we have to use evidence above at this node. Now we use Bayes rule, 
        {\rm{P(A11 | A15, A10)  =  }}\frac{{P(A15,{\rm{ A10 | A11) P(A11)}}}}{{{\rm{P(A15, A10)}}}}

Here, 
        {K_1}{\rm{  =  }}\frac{1}{{{\rm{P(A15, A10)}}}}
and P(M1) is given in the network. 

The algorithm. stops with, 
        P(A15, A10 | A11.) = P(A15 | A10, A11) P(A10 | A11) 
        
        P(A15, A10 | A11.) = P(A15 | A10, A11) P(A10) 

as A10 and Al1 are independent of each other. 

Finally all the results are combined. and K and K1 are calculate to find the solution to 
        P(B | Al2, A13, A14, A11). 

Evidence above and evidence below algorithms have a linear complexity in the number of nodes present in the network. 

Evidence Above and Below 

If we have evidence above 'B' and below 'B' like, 
                P(B | {A5, A4}, {Al2, A13, A14, A11})

then the evidences are separated into above denoted by  {\varepsilon ^ + }  and below as {\varepsilon ^ - }.

Now, Bayes rule is used to write, 
                P(B|{\varepsilon ^ + },{\varepsilon ^ - }) = \frac{{P({\varepsilon ^ - }|B,{\varepsilon ^ + })P(B|{\varepsilon ^ - })}}{{P({\varepsilon ^ - }|{\varepsilon ^ + })}}

Now, we use  K{}_2 = \frac{1}{{P({\varepsilon ^ - }|{\varepsilon ^ + })}} as a factor of normalization. 

The equation becomes, 
                P(B|{\varepsilon ^ + },{\varepsilon ^ - }) = {K_2}P({\varepsilon ^ - }|B,{\varepsilon ^ + })P(B|{\varepsilon ^ + }) 

In can be noted that and {\varepsilon ^ - } and {\varepsilon ^ + } are d-separated by B, so
                P(B|{\varepsilon ^ + },{\varepsilon ^ - }) = {K_2}P({\varepsilon ^ - }|B)P(B|{\varepsilon ^ + })

These probabilities can be got from evidence above and evidence below and   P(B|{\varepsilon ^ + },{\varepsilon ^ - }) can be calculated. 

To Top