Conditional Independence Relation

From GM-RKB
Jump to: navigation, search

A Conditional Independence Relation is a statistical relation that is true if and only if [math]\Pr(A \mid B \cap C) = \Pr(A \mid C)[/math], where [math]Pr[/math] is a conditional probability function.



References

2011

  • http://en.wikipedia.org/wiki/Conditional_independence
    • In probability theory, two events [math]R[/math] and [math]B[/math] are conditionally independent given a third event [math]Y[/math] precisely if the occurrence or non-occurrence of [math]R[/math] and the occurrence or non-occurrence of [math]B[/math] are independent events in their conditional probability distribution given Y. In other words, [math]R[/math] and [math]B[/math] are conditionally independent if and only if, given knowledge of whether [math]Y[/math] occurs, knowledge of whether [math]R[/math] occurs provides no information on the likelihood of [math]B[/math] occurring, and knowledge of whether [math]B[/math] occurs provides no information on the likehood of [math]R[/math] occurring. In the standard notation of probability theory, [math]R[/math] and [math]B[/math] are conditionally independent given [math]Y[/math] if and only if [math]\Pr(R \cap B \mid Y) = \Pr(R \mid Y)\Pr(B \mid Y),\,[/math] or equivalently, [math]\Pr(R \mid B \cap Y) = \Pr(R \mid Y).\,[/math] Two random variables [math]X[/math] and [math]Y[/math] are conditionally independent given a third random variable [math]Z[/math] if and only if they are independent in their conditional probability distribution given Z. That is, [math]X[/math] and [math]Y[/math] are conditionally independent given [math]Z[/math] if and only if, given any value of [math]Z[/math], the probability distribution of [math]X[/math] is the same for all values of [math]Y[/math] and the probability distribution of [math]Y[/math] is the same for all values of X.

2005