-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathindex.tex
More file actions
2595 lines (2123 loc) · 109 KB
/
index.tex
File metadata and controls
2595 lines (2123 loc) · 109 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
% Options for packages loaded elsewhere
% Options for packages loaded elsewhere
\PassOptionsToPackage{unicode}{hyperref}
\PassOptionsToPackage{hyphens}{url}
\PassOptionsToPackage{dvipsnames,svgnames,x11names}{xcolor}
%
\documentclass[
letterpaper,
DIV=11,
numbers=noendperiod]{scrreprt}
\usepackage{xcolor}
\usepackage{amsmath,amssymb}
\setcounter{secnumdepth}{5}
\usepackage{iftex}
\ifPDFTeX
\usepackage[T1]{fontenc}
\usepackage[utf8]{inputenc}
\usepackage{textcomp} % provide euro and other symbols
\else % if luatex or xetex
\usepackage{unicode-math} % this also loads fontspec
\defaultfontfeatures{Scale=MatchLowercase}
\defaultfontfeatures[\rmfamily]{Ligatures=TeX,Scale=1}
\fi
\usepackage{lmodern}
\ifPDFTeX\else
% xetex/luatex font selection
\fi
% Use upquote if available, for straight quotes in verbatim environments
\IfFileExists{upquote.sty}{\usepackage{upquote}}{}
\IfFileExists{microtype.sty}{% use microtype if available
\usepackage[]{microtype}
\UseMicrotypeSet[protrusion]{basicmath} % disable protrusion for tt fonts
}{}
\makeatletter
\@ifundefined{KOMAClassName}{% if non-KOMA class
\IfFileExists{parskip.sty}{%
\usepackage{parskip}
}{% else
\setlength{\parindent}{0pt}
\setlength{\parskip}{6pt plus 2pt minus 1pt}}
}{% if KOMA class
\KOMAoptions{parskip=half}}
\makeatother
% Make \paragraph and \subparagraph free-standing
\makeatletter
\ifx\paragraph\undefined\else
\let\oldparagraph\paragraph
\renewcommand{\paragraph}{
\@ifstar
\xxxParagraphStar
\xxxParagraphNoStar
}
\newcommand{\xxxParagraphStar}[1]{\oldparagraph*{#1}\mbox{}}
\newcommand{\xxxParagraphNoStar}[1]{\oldparagraph{#1}\mbox{}}
\fi
\ifx\subparagraph\undefined\else
\let\oldsubparagraph\subparagraph
\renewcommand{\subparagraph}{
\@ifstar
\xxxSubParagraphStar
\xxxSubParagraphNoStar
}
\newcommand{\xxxSubParagraphStar}[1]{\oldsubparagraph*{#1}\mbox{}}
\newcommand{\xxxSubParagraphNoStar}[1]{\oldsubparagraph{#1}\mbox{}}
\fi
\makeatother
\usepackage{color}
\usepackage{fancyvrb}
\newcommand{\VerbBar}{|}
\newcommand{\VERB}{\Verb[commandchars=\\\{\}]}
\DefineVerbatimEnvironment{Highlighting}{Verbatim}{commandchars=\\\{\}}
% Add ',fontsize=\small' for more characters per line
\usepackage{framed}
\definecolor{shadecolor}{RGB}{241,243,245}
\newenvironment{Shaded}{\begin{snugshade}}{\end{snugshade}}
\newcommand{\AlertTok}[1]{\textcolor[rgb]{0.68,0.00,0.00}{#1}}
\newcommand{\AnnotationTok}[1]{\textcolor[rgb]{0.37,0.37,0.37}{#1}}
\newcommand{\AttributeTok}[1]{\textcolor[rgb]{0.40,0.45,0.13}{#1}}
\newcommand{\BaseNTok}[1]{\textcolor[rgb]{0.68,0.00,0.00}{#1}}
\newcommand{\BuiltInTok}[1]{\textcolor[rgb]{0.00,0.23,0.31}{#1}}
\newcommand{\CharTok}[1]{\textcolor[rgb]{0.13,0.47,0.30}{#1}}
\newcommand{\CommentTok}[1]{\textcolor[rgb]{0.37,0.37,0.37}{#1}}
\newcommand{\CommentVarTok}[1]{\textcolor[rgb]{0.37,0.37,0.37}{\textit{#1}}}
\newcommand{\ConstantTok}[1]{\textcolor[rgb]{0.56,0.35,0.01}{#1}}
\newcommand{\ControlFlowTok}[1]{\textcolor[rgb]{0.00,0.23,0.31}{\textbf{#1}}}
\newcommand{\DataTypeTok}[1]{\textcolor[rgb]{0.68,0.00,0.00}{#1}}
\newcommand{\DecValTok}[1]{\textcolor[rgb]{0.68,0.00,0.00}{#1}}
\newcommand{\DocumentationTok}[1]{\textcolor[rgb]{0.37,0.37,0.37}{\textit{#1}}}
\newcommand{\ErrorTok}[1]{\textcolor[rgb]{0.68,0.00,0.00}{#1}}
\newcommand{\ExtensionTok}[1]{\textcolor[rgb]{0.00,0.23,0.31}{#1}}
\newcommand{\FloatTok}[1]{\textcolor[rgb]{0.68,0.00,0.00}{#1}}
\newcommand{\FunctionTok}[1]{\textcolor[rgb]{0.28,0.35,0.67}{#1}}
\newcommand{\ImportTok}[1]{\textcolor[rgb]{0.00,0.46,0.62}{#1}}
\newcommand{\InformationTok}[1]{\textcolor[rgb]{0.37,0.37,0.37}{#1}}
\newcommand{\KeywordTok}[1]{\textcolor[rgb]{0.00,0.23,0.31}{\textbf{#1}}}
\newcommand{\NormalTok}[1]{\textcolor[rgb]{0.00,0.23,0.31}{#1}}
\newcommand{\OperatorTok}[1]{\textcolor[rgb]{0.37,0.37,0.37}{#1}}
\newcommand{\OtherTok}[1]{\textcolor[rgb]{0.00,0.23,0.31}{#1}}
\newcommand{\PreprocessorTok}[1]{\textcolor[rgb]{0.68,0.00,0.00}{#1}}
\newcommand{\RegionMarkerTok}[1]{\textcolor[rgb]{0.00,0.23,0.31}{#1}}
\newcommand{\SpecialCharTok}[1]{\textcolor[rgb]{0.37,0.37,0.37}{#1}}
\newcommand{\SpecialStringTok}[1]{\textcolor[rgb]{0.13,0.47,0.30}{#1}}
\newcommand{\StringTok}[1]{\textcolor[rgb]{0.13,0.47,0.30}{#1}}
\newcommand{\VariableTok}[1]{\textcolor[rgb]{0.07,0.07,0.07}{#1}}
\newcommand{\VerbatimStringTok}[1]{\textcolor[rgb]{0.13,0.47,0.30}{#1}}
\newcommand{\WarningTok}[1]{\textcolor[rgb]{0.37,0.37,0.37}{\textit{#1}}}
\usepackage{longtable,booktabs,array}
\usepackage{calc} % for calculating minipage widths
% Correct order of tables after \paragraph or \subparagraph
\usepackage{etoolbox}
\makeatletter
\patchcmd\longtable{\par}{\if@noskipsec\mbox{}\fi\par}{}{}
\makeatother
% Allow footnotes in longtable head/foot
\IfFileExists{footnotehyper.sty}{\usepackage{footnotehyper}}{\usepackage{footnote}}
\makesavenoteenv{longtable}
\usepackage{graphicx}
\makeatletter
\newsavebox\pandoc@box
\newcommand*\pandocbounded[1]{% scales image to fit in text height/width
\sbox\pandoc@box{#1}%
\Gscale@div\@tempa{\textheight}{\dimexpr\ht\pandoc@box+\dp\pandoc@box\relax}%
\Gscale@div\@tempb{\linewidth}{\wd\pandoc@box}%
\ifdim\@tempb\p@<\@tempa\p@\let\@tempa\@tempb\fi% select the smaller of both
\ifdim\@tempa\p@<\p@\scalebox{\@tempa}{\usebox\pandoc@box}%
\else\usebox{\pandoc@box}%
\fi%
}
% Set default figure placement to htbp
\def\fps@figure{htbp}
\makeatother
\setlength{\emergencystretch}{3em} % prevent overfull lines
\providecommand{\tightlist}{%
\setlength{\itemsep}{0pt}\setlength{\parskip}{0pt}}
\KOMAoption{captions}{tableheading}
\makeatletter
\@ifpackageloaded{bookmark}{}{\usepackage{bookmark}}
\makeatother
\makeatletter
\@ifpackageloaded{caption}{}{\usepackage{caption}}
\AtBeginDocument{%
\ifdefined\contentsname
\renewcommand*\contentsname{Table of contents}
\else
\newcommand\contentsname{Table of contents}
\fi
\ifdefined\listfigurename
\renewcommand*\listfigurename{List of Figures}
\else
\newcommand\listfigurename{List of Figures}
\fi
\ifdefined\listtablename
\renewcommand*\listtablename{List of Tables}
\else
\newcommand\listtablename{List of Tables}
\fi
\ifdefined\figurename
\renewcommand*\figurename{Figure}
\else
\newcommand\figurename{Figure}
\fi
\ifdefined\tablename
\renewcommand*\tablename{Table}
\else
\newcommand\tablename{Table}
\fi
}
\@ifpackageloaded{float}{}{\usepackage{float}}
\floatstyle{ruled}
\@ifundefined{c@chapter}{\newfloat{codelisting}{h}{lop}}{\newfloat{codelisting}{h}{lop}[chapter]}
\floatname{codelisting}{Listing}
\newcommand*\listoflistings{\listof{codelisting}{List of Listings}}
\usepackage{amsthm}
\theoremstyle{remark}
\AtBeginDocument{\renewcommand*{\proofname}{Proof}}
\newtheorem*{remark}{Remark}
\newtheorem*{solution}{Solution}
\newtheorem{refremark}{Remark}[chapter]
\newtheorem{refsolution}{Solution}[chapter]
\makeatother
\makeatletter
\makeatother
\makeatletter
\@ifpackageloaded{caption}{}{\usepackage{caption}}
\@ifpackageloaded{subcaption}{}{\usepackage{subcaption}}
\makeatother
\usepackage{bookmark}
\IfFileExists{xurl.sty}{\usepackage{xurl}}{} % add URL line breaks if available
\urlstyle{same}
\hypersetup{
pdftitle={Probability and Statistics for ISP},
pdfauthor={T. A. Severini},
colorlinks=true,
linkcolor={blue},
filecolor={Maroon},
citecolor={Blue},
urlcolor={Blue},
pdfcreator={LaTeX via pandoc}}
\title{Probability and Statistics for ISP}
\usepackage{etoolbox}
\makeatletter
\providecommand{\subtitle}[1]{% add subtitle to \maketitle
\apptocmd{\@title}{\par {\large #1 \par}}{}{}
}
\makeatother
\subtitle{Course Notes for Statistics 383}
\author{T. A. Severini}
\date{2025-12-31}
\begin{document}
\maketitle
\renewcommand*\contentsname{Table of contents}
{
\hypersetup{linkcolor=}
\setcounter{tocdepth}{2}
\tableofcontents
}
\bookmarksetup{startatroot}
\chapter*{Preface}\label{preface}
\addcontentsline{toc}{chapter}{Preface}
\markboth{Preface}{Preface}
This is a Quarto book.
To learn more about Quarto books visit
\url{https://quarto.org/docs/books}.
\begin{Shaded}
\begin{Highlighting}[]
\DecValTok{1} \SpecialCharTok{+} \DecValTok{1}
\end{Highlighting}
\end{Shaded}
\begin{verbatim}
[1] 2
\end{verbatim}
\bookmarksetup{startatroot}
\chapter{Probability}\label{sec-chap1}
\newcommand{\cip}{\xrightarrow[]{p}}
\renewcommand{\P}{\mbox{P}}
\newcommand{\E}{\mbox{E}}
\newcommand{\Fhat}{\hat{F}}
\newcommand{\intii}{\int_{-\infty}^\infty}
\newcommand{\Ybar}{\bar{Y}}
\newcommand{\Var}{\mbox{Var}}
\newcommand{\lam}{\lambda}
\newcommand{\Ph}{\widehat{\mbox{P}}}
\newcommand{\I}{\mbox{I}}
\newcommand{\cid}{\xrightarrow[]{D}}
\newcommand{\bfmu}{\bm{\mu}}
\newcommand{\bfw}{\bm{w}}
\newcommand{\bfR}{\bm{R}}
\newcommand{\bfone}{\bm{1}}
\newcommand{\bfx}{\bm{x}}
\newcommand{\bfy}{\bm{y}}
\newcommand{\bfz}{\bm{z}}
\newcommand{\bfV}{\bm{\mathcal{V}}}
\newcommand{\bfVm}{\bm{V}}
\newcommand{\bfq}{\bm{q}}
\newcommand{\bfC}{\bm{C}}
\newcommand{\bfI}{\bm{I}}
\newcommand{\bfdelta}{\bm{\delta}}
\newcommand{\bfbeta}{\bm{\beta}}
\newcommand{\bfS}{\bm{S}}
\newcommand{\bfalpha}{\bm{\alpha}}
\newcommand{\bfepsilon}{\bm{\epsilonilon}}
\newcommand{\bfd}{\bm{d}}
\newcommand{\bfSigma}{\bm{\Sigma}}
\newcommand{\tbfSigma}{\tilde{\bfSigma}}
\newcommand{\bfzero}{\bm{0}}
\newcommand{\bfA}{\bm{A}}
\newcommand{\bfa}{\bm{a}}
\newcommand{\bfB}{\bm{B}}
\newcommand{\bfh}{\bm{h}}
\newcommand{\bfv}{\bm{v}}
\newcommand{\bfX}{\bm{X}}
\newcommand{\bfu}{\bm{u}}
\newcommand{\bfg}{\bm{g}}
\newcommand{\bfb}{\bm{b}}
\newcommand{\muhat}{\hat{\mu}}
\newcommand{\bfM}{\bm{M}}
\newcommand{\bfGamma}{\bm{\Gamma}}
\newcommand{\p}{{q}}
\newcommand{\bfF}{\bm{F}}
\newcommand{\phat}{\hat{p}}
\newcommand{\that}{\hat{\theta}}
\newcommand{\bbar}{\bar{b}}
\newcommand{\mhat}{\hat{m}}
\newcommand{\knumer}{\frac{1}{nh} \sum_{j=1}^n K(\frac{x - X_j}{h})Y_j }
\newcommand{\given}{{\ | \ }}
\newcommand{\fhat}{\hat{f}}
\newcommand{\Cov}{\mbox{Cov}}
\newcommand{\dmX}{\mathbb{{X}}}
\newcommand{\covS}{\mathbb{S}}
\newcommand{\bfe}{\bm{e}}
\newcommand{\bfell}{\bm{\ell}}
\newcommand{\bfY}{\bm{Y}}
\newcommand{\bfL}{\bm{L}}
\newcommand{\bfPsi}{\bm{\Psi}}
\newcommand{\bfQ}{\bm{Q}}
\newcommand{\bfT}{\bm{T}}
\newcommand{\X}{\mathcal{X}}
\newcommand{\Y}{\mathcal{Y}}
\newcommand{\xbar}{\bar{x}}
\newcommand{\theto}{\theta_0}
\newcommand{\bfZ}{\bm{Z}}
\newcommand{\bfP}{\bm{P}}
\newcommand{\bfD}{\bm{D}}
\newcommand{\bfO}{\bm{O}}
\newcommand{\bfLambda}{\bm{\Lambda}}
\newcommand{\bfc}{\bm{c}}
\newcommand{\bfE}{\bm{E}}
\newcommand{\bfU}{\bm{U}}
\newcommand{\bff}{\bm{f}}
\newcommand{\bfH}{\bm{H}}
\newcommand{\Xbar}{\bar{X}}
\newcommand{\ybar}{\bar{y}}
\newcommand{\calP}{\mathcal{P}}
\newcommand{\calY}{\mathcal{Y}}
\newcommand{\calX}{\mathcal{X}}
\newcommand{\calR}{\mathcal{R}}
\newcommand\BB{\rule[-1ex]{0pt}{0pt}}
\newcommand{\psihat}{{\widehat{\psi}}}
\newcommand{\thetast}{\theta^*}
\newcommand{\thetst}{\theta^*}
\newcommand\T{\rule{0pt}{2.6ex}}
\newcommand\B{\rule[-1ex]{0pt}{0pt}}
\section{Experiments}\label{experiments}
The starting point for probability theory is the concept of an
\emph{experiment}. The term experiment may actually refer to a physical
experiment in the usual sense, but more generally we will refer to
something as an experiment when it has the following properties:
\begin{itemize}
\item
There is a well-defined set of possible outcomes of the experiment;
\item
Each time the experiment is performed exactly one of the possible
outcomes occurs;
\item
The outcome that occurs is governed by some chance mechanism.
\end{itemize}
Let \(\Omega\) denote the \emph{sample space} of the experiment, the set
of possible outcomes of the experiment; the term \emph{outcome space} is
also used. We will refer to the elements of \(\Omega\) as \emph{basic
outcomes} and use the symbol \(\omega\) to denote a generic basic
outcome.
\phantomsection\label{onedie1}{} Consider the experiment in which we
roll a die. Then \[\Omega = \left\{1, 2, 3, 4, 5, 6 \right\}\] where,
for example, \(1\) denotes the outcome that we roll a \(1\).
Consider the experiment in which we choose a number from the interval
\((0, 1)\). Then \(\Omega = (0, 1)\).
\phantomsection\label{urn3}{} Suppose that we have an urn that contains
three balls, two red balls and one black ball. Consider the experiment
in which we successively choose two balls from the urn, that is, the
balls are chosen in such a way that we know which ball was chosen first.
Then the sample space for the experiment can be written
\[\Omega = \left\{ (R, R), \ (R, B), \ (B, R) \right\},\] where, for
example, \((R, B)\) means that the first ball selected is red and the
second ball selected is black.
Now suppose that the order in which the balls were selected is not
recorded. Then the sample space of the experiment is given by
\[\Omega = \left\{ \{R, R\}, \ \{R, B\} \right\},\] where, for example,
\(\{R, B\}\) means that one red ball is selected and one black ball is
selected.
\section{Events}\label{events}
Consider an experiment with sample space \(\Omega\). A subset \(A\) of
\(\Omega\) is called an \emph{event}. Let \(A\) be an event. Then, for
each \(\omega\in\Omega\), either \(\omega\in A\) or \(\omega\notin A\).
That is, when the experiment is performed, either \(A\) occurs (the
observed outcome is in \(A\)) or it doesn't occur (the observed outcome
is not in \(A\)).
\phantomsection\label{onedie2}{} Consider the experiment in which we
roll a die. Then \[A = \left\{2, 4, 6 \right\}\] is the event that we
roll an even number.
The event that we roll a number less than or equal to \(3\) is given by
\[B = \left\{1, 2, 3 \right\}.\]
The event that we roll a \(5\) is given by \[C = \left\{ 5 \right\}.\]
The event that we do not roll an even number, that is, that we roll an
odd number is given by \(A^c\), the complement of \(A\). Thus,
\[A^c = \left\{1, 3, 5 \right\}.\]
Consider the experiment in which two balls are drawn successively from
an urn containing two red balls and one black ball; the sample space for
this experiment is given in Example \hyperref[urn3]{{[}urn3{]}}.
Let \(A\) denote the event that exactly one red ball is selected. Then
\[A = \left\{ (R, B), \ (B, R) \right\}.\]
Because events are defined in terms of sets, sets play a central role in
probability theory. Here are few basic properties. Let \(A, B, C\) be
subsets of a sample space \(\Omega\); that is, let \(A, B, C\) be
events. Recall that \(A \cup B\), the \emph{union} of \(A\) and \(B\),
is the set consisting of all elements that are either in \(A\), in
\(B\), or in both \(A\) and \(B\); \(A\cap B\), the \emph{intersection}
of \(A\) and \(B\), is the set consisting of all elements that are in
both \(A\) and \(B\). Then
\begin{itemize}
\item
\((A \cup B)\cap C = (A \cap C) \cup (B \cap C)\)
\item
\((A \cap B) \cup C = (A \cup C) \cap (B \cup C)\)
\item
\((A\cup B)^c = A^c \cap B^c\)
\item
\((A \cap B)^c = A^c \cup B^c\).
\end{itemize}
If these properties are unfamiliar, you can show why they hold using
Venn diagrams. E.g., consider \((A \cup B)^c = A^c \cap B^c\). Figures
\hyperref[DeMorg1]{1.1} -- \hyperref[DeMorg3]{1.3} contain Venn diagrams
of \((A\cup B)^c\), \(A^c\), and \(B^c\), respectively. From these
diagrams, we can see that \((A \cup B)^c = A^c \cap B^c\).
\begin{figure}
{\centering \includegraphics[width=0.5\linewidth,height=\textheight,keepaspectratio]{demorgan1.pdf}
}
\caption{\((A \cup B)^c\)}
\end{figure}%
\begin{figure}
{\centering \includegraphics[width=0.5\linewidth,height=\textheight,keepaspectratio]{demorgan2new.pdf}
}
\caption{\(A^c\)}
\end{figure}%
\begin{figure}
{\centering \includegraphics[width=0.5\linewidth,height=\textheight,keepaspectratio]{demorgan3.pdf}
}
\caption{\(B^c\)}
\end{figure}%
\section{Probability Functions}\label{prob_fcns}
Consider an experiment with sample space \(\Omega\); recall that the
outcome of an experiment depends on some ``chance mechanism". It follows
that whether or not an event \(A\) occurs depends on that chance
mechanism and we use probability theory to describe the likelihood that
a given event occurs.
Therefore, associated with each event \(A\) is a probability
\(\mbox{P}(A)\). Here \(\mbox{P}\) is a function defined on subsets of
\(\Omega\) and taking values in the interval \([0, 1]\). The function
\(\mbox{P}\) is required to have certain properties:
\begin{itemize}
\item
\(\mbox{P}(\Omega) = 1\)
\item
If \(A\) and \(B\) are disjoint subsets of \(\Omega\), that is,
\(A\cap B = \emptyset\), then
\(\mbox{P}(A \cup B) = \mbox{P}(A) + \mbox{P}(B)\).
\item
If \(A_1, A_2, \ldots,\) are disjoint subsets of \(\Omega\), then
\[\mbox{P}( \cup_{n=1}^\infty A_n) = \sum_{n=1}^\infty \mbox{P}(A_n).\]
\end{itemize}
Note that when subsets of \(\Omega\) are disjoint, the corresponding
events are said to be \emph{mutually exclusive}.
\phantomsection\label{uni_ex}{} Suppose that \(\Omega = (0, 1)\) and
suppose that the probability of any interval in \(\Omega\) is the length
of the interval. More generally, we may take the probability of a subset
\(A\) of \(\Omega\) to be \[\mbox{P}(A) = \int_{A} dx.\]
For example, \[\mbox{P}\left( (0.2, 0.7) \right) = 0.5.\]
Consider the experiment of rolling one die, as discussed in Examples
\hyperref[onedie1]{{[}onedie1{]}} and \hyperref[onedie2]{{[}onedie2{]}}
and let \(\Omega\) denote the sample space of the experiment.
For \(A \subset \Omega\), let \(\mbox{P}(A) = |A|/6\), the number of
elements in \(A\), divided by \(6\).
E.g., the probability of rolling an even number is \(1/2\) and the
probability of rolling a number greater than or equal to \(5\) is
\(1/3\).
Note that, when an event consists of a single basic outcome \(\omega\),
we will write the probability of the event as \(\mbox{P}(\omega)\),
rather than as \(\mbox{P}(\left\{ \omega \right\})\), which is
technically correct (because the argument of the probability function
should be a set).
For instance, in the previous example, the probability of rolling a
\(6\) will be written as \(\mbox{P}(6)\) instead of as
\(\mbox{P}(\{ 6 \})\).
When \(\Omega\) is a countable set, then, by properties (P2) and (P3),
the probability of any event is given by the sum of the probabilities of
the basic outcomes corresponding to the event:
\[\mbox{P}(A) = \sum_{\omega \in A} \mbox{P}( \omega).\]
\phantomsection\label{binom_ex}{} Consider an experiment with sample
space
\[\Omega = \left\{(0, 0), \ (1, 0), \ (0, 1), \ (1, 1) \right\}.\]
For \(\omega = (x_1, x_2) \in \Omega\), take \[\begin{aligned}
\mbox{P}(\omega) &= \theta^{x_1} (1-\theta)^{1-x_1} \theta^{x_2} (1-\theta)^{1-x_2} \\
&= \theta^{x_1 + x_2} (1-\theta)^{2 - x_1 - x_2}
\end{aligned}\] where \(0 < \theta < 1\) is a given constant.
Thus, the four elements of \(\Omega\) have probabilities
\((1-\theta)^2, \theta(1-\theta), \theta(1-\theta), \theta^2\),
respectively.
Let \(A\) denote the event that exactly \(1\) one is observed; then
\[A = \left\{ (0, 1), \ (1, 0) \right\}.\] It follows that the
probability of \(A\) is the sum of the probabilities of the two basic
outcomes in \(A\):
\[\mbox{P}(A) = \mbox{P}\left( (1, 0) \cup (0, 1) \right) = \mbox{P}\left((1, 0) \right) + \mbox{P}\left( (0, 1) \right) = \theta(1-\theta) + \theta(1-\theta) = 2\theta(1-\theta).\]
\subsection*{Some implications of (P1) --
(P3)}\label{some-implications-of-p1-p3}
\addcontentsline{toc}{subsection}{Some implications of (P1) -- (P3)}
There are a number of straightforward consequences of properties
(P1)-(P3). For instance, because \(\Omega \cup \emptyset = \Omega\) and
\(\Omega \cap \emptyset = \emptyset\), by (P2)
\[\mbox{P}(\Omega) = \mbox{P}(\Omega) + \mbox{P}(\emptyset);\] it
follows that \(\mbox{P}(\emptyset) = 0\).
Let \(A^c\) denote the complement of an set \(A\subset \Omega\). Then
\(A \cup A^c = \Omega\) and \(A \cap A^c = \emptyset\). It follows from
(P2) that \[\mbox{P}(\Omega) = \mbox{P}(A) + \mbox{P}(A^c);\] it now
follows from (P1) that \[\mbox{P}(A^c) = 1 - \mbox{P}(A).\]
Suppose that \(A_1\) and \(A_2\) are subsets of \(\Omega\) that are not
necessarily disjoint. Then
\[\mbox{P}(A_1 \cup A_2) =\mbox{P}(A_1) + \mbox{P}(A_2) - \mbox{P}(A_1 \cap
A_2).\]
This important result is a little more difficult to prove than the
others we have considered. First note that \(A_1 \cup A_2\) can be
written as the union of three sets, \(A_1 \cap A_2\),
\(A_1 \cap A_2^c\), and \(A_1^c \cap A_2\). Furthermore, these three
sets are disjoint. An example of this fact is given in the Venn diagram
in Figure \hyperref[venn]{1.4}. In that diagram, the blue region is
\(A_1 \cap A_2^c\), the yellow region is \(A_1^c \cap A_2\), and the
brown region is \(A_1 \cap A_2\); combining these three regions forms
\(A_1 \cup A_2\).
\begin{figure}
{\centering \includegraphics[width=0.75\linewidth,height=\textheight,keepaspectratio]{vplot.pdf}
}
\caption{Venn Diagram Used to Illustrate
\(A_1 \cup A_2 = (A_1 \cap A_2) \cup (A_1 \cap A_2^c) \cup
(A_1^c \cap A_2)\)}
\end{figure}%
From the Venn diagram we can also see that
\[A_1 = (A_1 \cap A_2^c) \cup (A_1 \cap A_2) \ \ \text{ and } \ \
A_2 = (A_1^c \cap A_2) \cup (A_1 \cap A_2).\]
It follows from two applications of (P2) that \[\label{probeq1}
\mbox{P}(A_1 \cup A_2) = \mbox{P}(A_1 \cap A_2) + \mbox{P}(A_1 \cap A_2^c) + \mbox{P}(A_1^c \cap A_2)\]
and also that
\[\mbox{P}(A_1) = \mbox{P}(A_1 \cap A_2^c) + \mbox{P}(A_1 \cap A_2) \ \ \text{ and } \ \
\mbox{P}(A_2) = \mbox{P}(A_1^c \cap A_2) + \mbox{P}(A_1 \cap A_2).\]
From the last two equations, we see that
\[\mbox{P}(A_1 \cap A_2^c) = \mbox{P}(A_1) - \mbox{P}(A_1 \cap A_2) \ \ \text{ and } \ \
\mbox{P}(A_1^c \cap A_2) = \mbox{P}(A_2) - \mbox{P}(A_1 \cap A_2).\]
Substituting these expressions into the right-hand side of equation
\hyperref[probeq1]{{[}probeq1{]}}, it follows that \[\begin{aligned}
\mbox{P}(A_1 \cup A_2) &= \mbox{P}(A_1 \cap A_2) + \mbox{P}(A_1) - \mbox{P}(A_1 \cap A_2) + \mbox{P}(A_2) - \mbox{P}(A_1 \cap A_2) \\
&= \mbox{P}(A_1) + \mbox{P}(A_2) - \mbox{P}(A_1 \cap A_2).
\end{aligned}\]
Consider an experiment with sample space \(\Omega\) and events
\(A_1, A_2\). Let \(A_1 \setminus A_2\) denote the elements of \(A_1\)
that are not in \(A_2\).
Suppose that \(A_2\subset A_1\). Then
\[\mbox{P}(A_1 \setminus A_2) = \mbox{P}(A_1) - \mbox{P}(A_2).\]
In general, \[A_1 = (A_1 \cap A_2) \cup (A_1 \cap A_2^c) =
(A_1 \cap A_2) \cup (A_1 \setminus A_2).\] Note that
\[A_1\cap A_2 \ \ \text{ and } \ \ A_1 \cap A_2^c\] are disjoint and
that \[A_1 \cap A_2^c = A_1 \setminus A_2.\] Hence,
\[\mbox{P}(A_1) = \mbox{P}(A_1 \cap A_2) + \mbox{P}(A_1 \setminus A_2)\]
so that
\[\mbox{P}(A_1 \setminus A_2) = \mbox{P}(A_1) - \mbox{P}(A_1 \cap A_2).\]
\subsection*{Interpretation of
probability}\label{interpretation-of-probability}
\addcontentsline{toc}{subsection}{Interpretation of probability}
Although we have described the properties of a probability function,
nothing has been said about what the probability function is measuring.
In a mathematical sense, that is irrelevant -- a probability function is
defined by its properties and any function satisfying those properties
can be used to calculate a ``probability".
However, in order to better understand the mathematical results, and to
develop some intuition regarding probability theory, it is useful to
have some notion of what is meant by ``probability". Several different
interpretations of probability are used in applications. The most
common, and the one we will use here, is the interpretation of
probability as a''limiting relative frequency".
Consider an experiment with sample space \(\Omega\) and let \(A\) denote
an event. According to the limiting relative frequency interpretation of
probability, the statement that \(A\) has probability \(0.4\) (for
example), means that if the experiment is repeated a large number of
times then in about \(40\%\) of those experiments the event \(A\) will
occur.
More formally, let \(N_n(A)\) denote the number of times the event \(A\)
occurs in \(n\) repetitions of the experiment. Then
\[\mbox{P}(A) = \lim_{n\to\infty} \frac{N_n(A)}{n};\] the right-hand
side of this expression is often described as a ``limiting relative
frequency".
\section{Sampling from a Finite
Population}\label{sampling-from-a-finite-population}
A particularly simple, but useful, case occurs when the sample space of
the experiment, \(\Omega\), is a finite set and each \(\omega\in\Omega\)
has the same probability.
Write
\[\Omega = \left\{ \omega_1, \omega_2, \ldots, \omega_m \right\};\] note
that any subset of \(\Omega\) can be written as the union of sets of the
form \(\{ \omega_j \}\), which are disjoint. Let
\(c = \mbox{P}(\omega_j)\) denote the common value of the probability of
each element of \(\Omega\). Then, because \(\mbox{P}(\Omega) = 1\) and
\[\mbox{P}(\Omega) = \mbox{P}(\omega_1) + \mbox{P}(\omega_2) + \cdots + \mbox{P}(\omega_m) = m c\]
we must have \(c = 1/|\Omega|\) where \(|\Omega|\) denotes the
cardinality of \(\Omega\), that is, the number of elements in
\(\Omega\).
Furthermore, for any \(A \subset \Omega\),
\[\mbox{P}(A) = \sum_{\omega\in A} \mbox{P}(\omega) = {|A| \over |\Omega|}.\]
Thus, the problem of determining \(\mbox{P}(A)\) is essentially the
problem of counting the number of elements in the set \(A\) and the
number of elements in \(\Omega\). The subfield of mathematics concerned
with counting the number of elements in a set is known as
\emph{combinatorics}.
In some cases, such as the one in the following example, the counting
needed is relatively straightforward.
\phantomsection\label{dice1}{}
Consider the experiment of rolling \(2\) dice, one at a time. The sample
space of the experiment can be written
\[\Omega = \left\{ (1, 1), (1, 2), \ldots, (1, 6), (2, 1), \ldots, (2, 6), \ldots, (6, 1), \ldots, (6, 6) \right\}\]
so that it has \(36\) elements. Suppose the dice are ``fair", in the
sense that each element of \(\Omega\) is equally likely.
Let \(A\) denote the event that result of the dice rolling is
``doubles", i.e., the two numbers rolled are equal, and suppose that we
are interested in the probability of \(A\).
As noted previously, the sample space \(\Omega\) has \(36\) elements.
The event \(A\) has \(6\) elements, \((1, 1), (2, 2), \ldots, (6, 6)\).
Thus, the probability of rolling doubles is \(6/36 = 1/6\).
In other cases, the counting needed is more complicated and it is useful
to apply one of the many well-known results that are used to solve such
counting problems. Here we consider only a few simple ones.
\subsection*{Counting principle}\label{counting-principle}
\addcontentsline{toc}{subsection}{Counting principle}
Many results in combinatorics are based on the \emph{counting
principle}. Let \(A\) and \(B\) denote finite sets and let \(A\times B\)
denote the Cartesian product of \(A\) and \(B\), that is, the set of the
form \[A\times B = \left\{ (a, b): \ a\in A, \ b\in B \right\}.\] Then
\[|A\times B |= |A| \ |B|.\]
Thus, if a task can be completed in \(r\) stages and there are \(n_j\)
ways to complete the \(j\)th stage, \(j=1, 2, \ldots, r\), then there
are \[n_1\times n_2 \times \cdots \times n_r\] ways to complete the
task.
\subsection*{Permutations and
combinations}\label{permutations-and-combinations}
\addcontentsline{toc}{subsection}{Permutations and combinations}
A \emph{permutation} of \(n\) distinct objects is an ordering of them.
The possible permutations of \(a, b, c\) are
\[abc, \ bac, \ cab, \ acb, \ bca, \ cba.\]
The number of permutations of \(n\) distinct objects can be found using
the counting principle, breaking the problem into stages. Note that
there are \(n\) ways to choose the first object, \(n-1\) ways to choose
the second object, and so on, until there is only \(1\) way to choose
the last object. Hence, there are
\[n\cdot (n-1) \cdots (2)\cdot (1) = n!\] ways to order the \(n\)
objects. That is, there are \(n!\) possible permutations of \(n\)
distinct objects.
In some cases, we might be interested in ordered \emph{samples} from a
given set.
The possible ordered samples of size \(2\) from the set \(\{a, b, c \}\)
are \[(a, b),\ (a, c),\ (b, a),\ (b, c), \ (c, a), \ (c, b).\]
The number of possible ordered samples of size \(k\) from a set of \(n\)
distinct elements can be found using the counting principle. There are
\(n\) ways to choose the first element, \(n-1\) ways to choose the
second element, and so on. However, in contrast to permutations, here we
stop selecting elements after the \(k\)th selection. Therefore, there
\[n (n-1) \cdots (n-k+1)\] ordered samples of size \(k\) from a set of
\(n\) elements; note there are \(k\) terms in this product. The
expression \(n(n-1)\cdots (n-k+1)\) is often denoted by \((n)_k\); it
can also be written as \[(n)_k = \frac{n!}{(n-k)!}.\]
\phantomsection\label{urn0}{} Consider an urn containing \(5\) balls,
\(2\) of which are black and \(3\) of which are red. Suppose that \(2\)
balls are randomly selected from the urn, without replacement; that is,
after the first ball is selected, it is not returned to the urn for the
second selection. Thus, there are \((5)_2 = 5(4) = 20\) basic outcomes
in \(\Omega\); ``randomly selected" means that any ordered pair of \(2\)
balls is equally likely to be selected.
Let \(A\) denote the event that a red ball is selected followed by a
black ball. Find \(\mbox{P}(A)\). Because each basic outcome is assumed
to have the same probability, \[\mbox{P}(A) = \frac{|A|}{|\Omega|}.\]
Thus, we need to count the number of basic outcomes in of \(A\).
To find the number of basic outcomes in \(A\), we use the facts that
there are \(3\) ways to choose the first (red) ball and \(2\) ways to
choose the second (black) ball. Thus, there are \((3)_2=3(2) = 6\) basic
outcomes in \(A\). It follows that
\[\mbox{P}(A) = \frac{6}{20} = \frac{3}{10}.\]
Now suppose that we are interested in the possible \emph{combinations}
of \(k\) objects chosen from a set of \(n\) distinct objects. When
considering combinations, the order of the objects is irrelevant; we are
interested only in the set of \(k\) objects.
The possible combinations of \(2\) elements chosen from the set
\(\{a, b, c \}\) are given by \[\{a, b \}, \ \{a, c \}, \ \{b, c \}.\]
The number of possible combinations of \(k\) objects chosen from a set
of \(n\) distinct objects is denoted by \[{n \choose k},\] read as
``\(n\) choose \(k\)".
To find an expression for \(n\) choose \(k\), consider choosing an
\textbf{ordered} sample of size \(k\) from \(n\) distinct elements.
This can be done in two steps:
\begin{enumerate}
\def\labelenumi{\arabic{enumi}.}
\item
choose \(k\) elements from \(n\)
\item
order the \(k\) elements
\end{enumerate}
We know that
\begin{itemize}
\item
there are \((n)_k\) ordered samples of size \(k\) from \(n\) distinct
elements
\item
there are \(n\) choose \(k\) ways to choose \(k\) elements from \(n\)
\item
there are \(k!\) ways to order \(k\) elements
\end{itemize}
Therefore, we must have \[(n)_k = {n \choose k}\, k!\] so that
\[{n \choose k} = \frac{(n)_k}{k!} = \frac{n!}{k!\, (n-k)!}.\]
Terms of the form \({n \choose k}\) are often called the \emph{binomial
coefficients}, because of the binomial formula:
\[(x+y)^n = \sum_{j=0}^n {n \choose j} x^j y^{n-j}\] for all real
numbers \(x, y\) and all positive integers \(n\).
Consider the framework of Example \hyperref[urn0]{{[}urn0{]}}: there is
an urn with \(2\) black balls and \(3\) red balls and \(2\) balls are
randomly selected from the urn. Let \(B\) denote the event that \(2\)
red balls are selected. Find \(\mbox{P}(B)\).
The basic outcomes here are the sets of two balls selected from the urn.
Because each basic outcome is assumed to have the same probability,
\[\mbox{P}(B) = \frac{|B|}{|\Omega|}.\]
The number of elements in \(\Omega\) is the number of ways to select
\(2\) balls from the set of \(5\), given by \[{5 \choose 2} = 10.\] The
number of elements in \(B\) is the number of ways to choose \(2\) red
balls from the set of \(3\) red balls, given by \[{3 \choose 2} = 3.\]
Thus, \(\mbox{P}(B) = 3/10\).
Suppose that \(5\) cards are dealt from a well-shuffled deck of playing
cards. Recall that, in such a deck, there are \(52\) cards and each card
falls into one of four suits (\(13\) cards in each suit).
What is the probability that all \(5\) cards are of the same suit (i.e.,
a ``flush", in poker)?
There are \[{52 \choose 5}\] ways to choose \(5\) cards from a deck of
\(52\). To find the number of ways in which \(5\) cards can be chosen
from one suit, we can use the counting principle: there are \(4\) ways
to choose the suit and, given the suit, there are \[{13 \choose 5}\]
ways to choose the \(5\) cards from the suit. Therefore, there are
\[4 {13 \choose 5}\] ways to choose \(5\) cards from one suit.
It follows that the probability of being dealt \(5\) cards from one suit
is \[\begin{aligned}
\frac{4 {13 \choose 5}}{{52 \choose 5}} &= \frac{4 \frac{13!}{8!\, 5!}}{\frac{52!}{47!\, 5!} }\\
&= 4 \frac{(13)(12)(11)(10)(9)}{(52)(51)(50)(49)(48)} \\
&= 0.00198.
\end{aligned}\]
In some cases, it is easier to find the probability of an event \(A\) by
finding the probability of \(A^c\) and using the fact that
\(\mbox{P}(A) = 1 - \mbox{P}(A^c)\). In fact, this simple result often
converts a complicated problem into a relatively easy one.
Suppose that \(n\) cards are dealt from a well-shuffled deck of playing
cards. What is the probability that at least \(1\) face card is drawn?
We can calculate the probability of being dealt at least one face card
by calculating the probability of being dealt no face cards and then
subtracting that result from 1.
In a standard deck of cards, there are \(12\) face cards and \(40\)
non-face cards.
To be dealt \(n\) non-face cards, \(n\) cards must be chosen from the
\(40\) non-face cards. There are \[{40 \choose n}\] ways to do this.
Since there are \[{52 \choose n}\] ways to choose \(n\) cards from the
entire deck, the probability of being dealt no face cards is
\[{40 \choose n} \over {52 \choose n}\] and, hence, the probability of
being dealt at least one face card is
\[1 - {{40 \choose n} \over {52 \choose n}} = 1 - {40\cdot 39 \cdots (40-n+1) \over 52\cdot 51 \cdots (52-n+1)}.\]
This result holds for \(n\leq 40\); otherwise the probability is \(0\).
\section{Conditional Probability}\label{conditional-probability}
Consider the dice-rolling experiment discussed in Example
\hyperref[dice1]{{[}dice1{]}}: two dice are rolled, one at a time. The
sample space of the experiment, \(\Omega\), has \(36\) elements,
\[\Omega = \left\{ (1, 1), (1, 2), \ldots, (1, 6), (2, 1), \ldots, (2, 6), \ldots, (6, 1), \ldots, (6, 6) \right\}\]
and each element of \(\Omega\) is equally likely.
Let \(A\) denote the event result of the experiment includes at least
\(1\) six; then
\[A = \left\{ (1, 6), (2, 6), (3, 6), (4, 6), (5, 6), (6, 1), (6, 2), (6, 3), (6, 4), (6, 5), (6, 6) \right\},\]
which has \(11\) elements. Hence, \(\mbox{P}(A) = 11/36\).
Now suppose that we know that the sum of the dice is at least \(10\).
Based on this information, what is the probability that the result
includes at least \(1\) six?
Note that this probability cannot be described in terms of a single
event, because it includes the condition that the sum of the dice is at
least \(10\). It is an example of a \emph{conditional probability}.
Let \(B\) denote the event that the sum of the dice is at least \(10\);
then
\[B = \left\{ (4, 6), (5, 5), (5, 6), (6, 4), (6, 5), (6, 6) \right\}.\]
We want to find \(\mbox{P}(A\, |\, B)\), read as the ``conditional
probability of \(A\) given \(B\)". It is the probability that we roll at
least \(1\) six \textbf{given that} the sum of the dice is at least
\(10\).
There are \(6\) elements in \(B\); for \(5\) of these, there is at least
one six. Therefore, it is reasonable to expect that
\(\mbox{P}(A\, |\, B) = 5/6\).
This is, in fact, correct. The general formula for a conditional
probability is
\[\mbox{P}(A\, |\, B) = \frac{\mbox{P}(A \cap B)}{\mbox{P}(B)},\]
provided that \(\mbox{P}(B) > 0\). Note that \(A\cap B\) represents the
part of \(A\) that satisfies the condition \(B\).
In the example,
\[A \cap B = \left\{ (4, 6), (5, 6), (6, 4), (6, 5), (6, 6) \right\}\]
so that \(\mbox{P}(A \cap B) = 5/36\). Using the fact that
\(\mbox{P}(B) = 6/36\) yields the answer given above.
Conditional probabilities are useful because they allow us to
incorporate additional information into the probability calculation.
Note that, for a given event \(B\) with \(\mbox{P}(B)>0\), the function
\(Q\) defined on subsets of \(\Omega\) and given by
\[\mbox{Q}(A) = \mbox{P}(A\, | \, B)\] is a probability function on
\(\Omega\), in the sense that it satisfies all the properties of a
probability function, such as
\[\mbox{P}(A^c\, | \, B) = 1 - \mbox{P}(A\, | \, B)\] and
\[\mbox{P}(A_1 \cup A_2\, |\, B) = \mbox{P}(A_1\, |\, B) + \mbox{P}(A_2\, |\, B) - \mbox{P}(A_1 \cap A_2\, | \, B).\]
Consider an urn with \(2\) red balls and \(w\) white balls for some
\(w\geq 2\). Suppose that \(2\) balls are randomly selected from the
urn. Given that the balls are the same color, what is the probability
that they are red?
Define two events, \(A\), the event that both balls are red and \(B\),
the event that the balls are the same color. Hence, we want to determine
\(\mbox{P}(A\, |\, B)\).
There are \[{2 + w \choose 2}\] ways to choose \(2\) balls from the urn,
which contains \(2+w\) balls. There is \(1\) way to choose \(2\) red
balls and \[{w \choose 2}\] ways to choose \(2\) white balls. Hence,
\[\mbox{P}(B) = \frac{1 + {w \choose 2}}{{2 + w \choose 2}} = \frac{w(w-1) + 2}{(w+2)(w+1)}\]
and
\[\mbox{P}(B) = \frac{1}{{2 + w \choose 2}} = \frac{2}{(w+2)(w+1)}.\]
Note that, because \(A\subset B\), \(A \cup B = A\). It follows that
\[\begin{aligned}
\mbox{P}(A\, |\, B) &= \frac{\mbox{P}(A \cap B)}{\mbox{P}(B)} = \frac{\mbox{P}(A)}{\mbox{P}(B)} \\
&= \frac{ \frac{2}{(w+2)(w+1)} }{ \frac{w(w-1) + 2}{(w+2)(w+1)}} \\
&= \frac{2}{w(w-1) + 2}.
\end{aligned}\]
\subsection*{Multiplication law}\label{multiplication-law}
\addcontentsline{toc}{subsection}{Multiplication law}
Rewriting the expression for conditional probability yields the
\emph{multiplication law} for probabilities: for events \(A\), \(B\),
\[\mbox{P}(A \cap B) = \mbox{P}(B\, | \, A)\mbox{P}(A) = \mbox{P}(A\, | \, B)\mbox{P}(B).\]
\phantomsection\label{urn1}{} Consider an urn with \(r\) red balls and
\(b\) black balls, where \(r\) and \(b\) are positive integers. Suppose
that \(2\) balls are randomly selected from the urn, one at a time. What
is the probability that the first ball is red and the second ball is
black?
Define two events, \(A\), the event that the first ball is red and
\(B\), the event that the second ball is black. We want
\(\mbox{P}(A \cap B)\).
This is a case in which the expression
\(\mbox{P}(B\, |\, A)\mbox{P}(A)\) may be a convenient way to calculate
\(\mbox{P}(A \cap B)\): the probability that the first ball is red is
easy to determine, \[\mbox{P}(A) = \frac{r}{r + b},\] and, given the
result on the first ball, the probability that the second ball is black
is also easy to determine.
Specifically, if the first ball is red, that leaves \(r-1\) red balls
and \(b\) black balls in the urn. Hence,
\[\mbox{P}(B\, | \, A) = \frac{b}{r -1 + b}.\] It follows that