Statistical Independence Relationship

From GM-RKB
(Redirected from statistical independence)
Jump to: navigation, search

A Statistical Independence Relationship is a binary relationship that determines whether the Probability Distribution of one Random Variable is unrelated to the Probability Distribution of the other Random Variable.



References

2015

  • (Wikipedia, 2015) ⇒ http://en.wikipedia.org/wiki/Independence_(probability_theory) Retrieved:2015-6-28.
    • In probability theory, two events are independent, statistically independent, or stochastically independent[1] if the occurrence of one does not affect the probability of the other. Similarly, two random variables are independent if the realization of one does not affect the probability distribution of the other.

      The concept of independence extends to dealing with collections of more than two events or random variables, in which case the events are pairwise independent if each pair are independent of each other, and the events are mutually independent if each event is independent of each other combination of events.

  1. Cite error: Invalid <ref> tag; no text was provided for refs named Artificial Intelligence

2009

  • (Wikipedia, 2009) ⇒ http://en.wikipedia.org/wiki/Statistical_independence
    • In probability theory, to say that two events are independent intuitively means that the occurrence of one event makes it neither more nor less probable that the other occurs. For example:
      • The event of getting a 6 the first time a die is rolled and the event of getting a 6 the second time are independent.
      • By contrast, the event of getting a 6 the first time a die is rolled and the event that the sum of the numbers seen on the first and second trials is 8 are dependent.
      • If two cards are drawn with replacement from a deck of cards, the event of drawing a red card on the first trial and that of drawing a red card on the second trial are independent.
      • By contrast, if two cards are drawn without replacement from a deck of cards, the event of drawing a red card on the first trial and that of drawing a red card on the second trial are dependent.
    • Similarly, two random variables are independent if the conditional probability distribution of either given the observed value of the other is the same as if the other's value had not been observed. The concept of independence extends to dealing with collections of more than two events or random variables.

2006

  • Suzanne R. Dubnicka. (2006). “STAT 510: Handout 1 - Probability Terminology. Kansas State University
    • TERMINOLOGY : We call two events A and B mutually exclusive, or disjoint, if A \ B = ; so that they have no outcomes in common. Thus, if A occurs then B cannot occur. Extending this definition to a finite or countable collection of sets is obvious.
  • Suzanne R. Dubnicka. (2006). “STAT 510: Handout 2 - Counting Techniques and More Probabililty. Kansas State University
    • TERMINOLOGY : When the occurrence or non-occurrence of A has no effect on whether or not B occurs, and vice-versa, we say that the events A and B are independent. Mathematically, we define A and B to be independent iff (if and only if) :[math] P(A ∩ B) = P(A)P(B).[/math] Otherwise, A and B are called dependent events. Note that if A and B are independent, [math] P(A|B) = P(A ∩ B) / P(B) = P(A)P(B)/P(B) = P(A)[/math] and [math]P(B|A) = P(B ∩ A) / P(A) = P(B)P(A) / P(A) = P(B).[/math]