-
Notifications
You must be signed in to change notification settings - Fork 2
/
Copy pathprobability.tex
142 lines (100 loc) · 3.23 KB
/
probability.tex
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
\renewcommand{\Pr}{\text{Pr}}
\newcommand{\Var}{\text{Var}}
\chapter{Probability}
Before we can discuss probablistic and randomized algorithms, we must
first develop the tools necessary to rigorously discuss probability.
We are concerned with the probability of sets of events, and for
simplicity we will simply call these sets events.
The probability of a event $A$ is $\Pr(A)$. The probability of an
event $A$ given that event $B$ has occured is $\Pr(A | B)$.
The probability of events $A$ and $B$ both happening is
\begin{center}
\begin{math}
\Pr(A \cap B)
= \Pr(A|B)\Pr(B) = \Pr(A)\Pr(B|A)
\end{math}
\end{center}
which is the definition of \emph{conjunctive probability}.
The probability of either events $A$ or $B$ happening is
\begin{center}
\begin{math}
\Pr(A \cup B)
= \Pr(A) + \Pr(B) - \Pr(A \cap B)
\end{math}
\end{center}
which is the definition of \emph{disjunctive probability}.
If the outcome of event $A$ does not affect that of $B$, we say $A$
and $B$ are \emph{independent}.
If we have a set of independent events $S$, from the definition of
disjunctive probability we get
\begin{displaymath}
%
\Pr \left( \bigcup_{s \in S} s \right) = \sum_{s \in S} \Pr(s)
%
\end{displaymath}
which is the \emph{linearity of independent probability}.
\section{Linearity of Expectation}
Often, we are interested in the outcome of some event with intrinsic
value, such as the result of rolling a die. In this case, we may be
interested in the average value. Since not all values are equally
likely, it is important to weight each value by its probability. We
call such a weighted average the \emph{Expected Value}.
The expected value of a random variable $X$ is
\begin{displaymath}
%
E(X) = \sum_{x \in X} x \cdot \Pr(x)
%
\end{displaymath}
which is the definition of \emph{expectation}.
If we have two events $X$ and $Y$, we have
\begin{center}
\begin{math}
E(X + Y) = E(X) + E(Y)
\end{math}
\end{center}
which is the \emph{linearity of expectation}.
\section{Markov's Inequality}
We often wish to establish bounds on probability, for which one
important result is for any random variable $X$ and $a > 0$
\begin{center}
\begin{math}
\Pr(X \geq a) \leq \frac{E(X)}{a}
\end{math}
\end{center}
which is \emph{Markov's Inequality}.
\section{Variance}
Given a random variable $X$ we have
\begin{center}
\begin{math}
\Var (X) = E(X^2) - E(X)^2
\end{math}
\end{center}
which is the definition of \emph{variance}.
If we have an independent and identically distributed set of events
$X$ then
\begin{displaymath}
\Var (\sum_{x \in X} x) = \sum_{x \in X} \Var (x)
\end{displaymath}
\section{Chebyshev's Inequality}
For any random variable $X$ and $a > 0$ we have
\begin{center}
\begin{math}
\Pr ( | X - E(X) | \geq a ) \leq \frac{\Var{X}}{a^2}
\end{math}
\end{center}
which is \emph{Chebyshev's Inequality}.
\section{Chernoff Bounds}
Where $X$ is the sum of indicator random variables, for any $ \epsilon
> 0 $ we have
\begin{center}
\begin{math}
\Pr ( X \geq ( 1 + \epsilon ) E(X) ) \leq e^{ \frac{-\epsilon^2 E(X)}{3}}
\end{math}
\end{center}
and
\begin{center}
\begin{math}
\Pr ( X \geq ( 1 - \epsilon ) E(X) ) \leq e^{ \frac{-\epsilon^2 E(X)}{2}}
\end{math}
\end{center}
which are \emph{Chernoff Bounds}.