MIT RES.6-012 Introduction to Probability, Spring 2018
4.0
(3)
24 learners
What you'll learn
This course includes
- 29.5 hours of video
- Certificate of completion
- Access on mobile and TV
Course content
1 modules • 266 lessons • 29.5 hours of video
MIT RES.6-012 Introduction to Probability, Spring 2018
266 lessons
• 29.5 hours
MIT RES.6-012 Introduction to Probability, Spring 2018
266 lessons
• 29.5 hours
- L01.1 Lecture Overview01:52
- L01.2 Sample Space05:38
- L01.3 Sample Space Examples05:03
- L01.4 Probability Axioms08:55
- L01.5 Simple Properties of Probabilities11:05
- L01.6 More Properties of Probabilities08:40
- L01.7 A Discrete Example05:13
- L01.8 A Continuous Example05:20
- L01.9 Countable Additivity12:10
- L01.10 Interpretations & Uses of Probabilities03:48
- S01.0 Mathematical Background Overview01:25
- S01.1 Sets10:55
- S01.2 De Morgan's Laws04:53
- S01.3 Sequences and their Limits06:00
- S01.4 When Does a Sequence Converge02:46
- S01.5 Infinite Series03:11
- S01.6 The Geometric Series04:07
- S01.7 About the Order of Summation in Series with Multiple Indices10:05
- S01.8 Countable and Uncountable Sets06:19
- S01.9 Proof That a Set of Real Numbers is Uncountable04:02
- S01.10 Bonferroni's Inequality09:28
- L02.1 Lecture Overview02:07
- L02.2 Conditional Probabilities09:00
- L02.3 A Die Roll Example05:02
- L02.4 Conditional Probabilities Obey the Same Axioms07:45
- L02.5 A Radar Example and Three Basic Tools10:59
- L02.6 The Multiplication Rule06:17
- L02.7 Total Probability Theorem05:25
- L02.8 Bayes' Rule04:28
- L03.1 Lecture Overview01:27
- L03.2 A Coin Tossing Example07:59
- L03.3 Independence of Two Events06:10
- L03.4 Independence of Event Complements02:59
- L03.5 Conditional Independence02:46
- L03.6 Independence Versus Conditional Independence05:30
- L03.7 Independence of a Collection of Events06:00
- L03.8 Independence Versus Pairwise Independence08:35
- L03.9 Reliability07:28
- L03.10 The King's Sibling06:54
- L04.1 Lecture Overview02:29
- L04.2 The Counting Principle11:12
- L04.3 Die Roll Example04:39
- L04.4 Combinations10:08
- L04.5 Binomial Probabilities06:38
- L04.6 A Coin Tossing Example11:48
- L04.7 Partitions05:20
- L04.8 Each Person Gets An Ace09:45
- L04.9 Multinomial Probabilities10:36
- L05.1 Lecture Overview01:40
- L05.2 Definition of Random Variables09:14
- L05.3 Probability Mass Functions10:21
- L05.4 Bernoulli & Indicator Random Variables03:06
- L05.5 Uniform Random Variables04:06
- L05.6 Binomial Random Variables06:08
- L05.7 Geometric Random Variables07:37
- L05.8 Expectation10:38
- L05.9 Elementary Properties of Expectation04:12
- L05.10 The Expected Value Rule10:00
- L05.11 Linearity of Expectations03:59
- S05.1 Supplement: Functions08:08
- L06.1 Lecture Overview02:02
- L06.2 Variance10:43
- L06.3 The Variance of the Bernoulli & The Uniform08:40
- L06.4 Conditional PMFs & Expectations Given an Event07:31
- L06.5 Total Expectation Theorem06:28
- L06.6 Geometric PMF Memorylessness & Expectation10:29
- L06.7 Joint PMFs and the Expected Value Rule10:16
- L06.8 Linearity of Expectations & The Mean of the Binomial08:25
- L07.1 Lecture Overview01:50
- L07.2 Conditional PMFs10:48
- L07.3 Conditional Expectation & the Total Expectation Theorem06:10
- L07.4 Independence of Random Variables05:08
- L07.5 Example04:44
- L07.6 Independence & Expectations04:22
- L07.7 Independence, Variances & the Binomial Variance07:09
- L07.8 The Hat Problem16:09
- S07.1 The Inclusion-Exclusion Formula11:13
- S07.2 The Variance of the Geometric05:42
- S07.3 Independence of Random Variables Versus Independence of Events06:51
- L08.1 Lecture Overview01:13
- L08.2 Probability Density Functions11:09
- L08.3 Uniform & Piecewise Constant PDFs02:52
- L08.4 Means & Variances06:57
- L08.5 Mean & Variance of the Uniform03:56
- L08.6 Exponential Random Variables08:09
- L08.7 Cumulative Distribution Functions12:48
- L08.8 Normal Random Variables09:14
- L08.9 Calculation of Normal Probabilities10:11
- L09.1 Lecture Overview01:33
- L09.2 Conditioning A Continuous Random Variable on an Event09:56
- L09.3 Conditioning Example03:08
- L09.4 Memorylessness of the Exponential PDF08:18
- L09.5 Total Probability & Expectation Theorems06:51
- L09.6 Mixed Random Variables05:35
- L09.7 Joint PDFs09:18
- L09.8 From The Joint to the Marginal07:23
- L09.9 Continuous Analogs of Various Properties01:40
- L09.10 Joint CDFs04:16
- S09.1 Buffon's Needle & Monte Carlo Simulation16:12
- L10.1 Lecture Overview01:42
- L10.2 Conditional PDFs06:57
- L10.3 Comments on Conditional PDFs04:34
- L10.4 Total Probability & Total Expectation Theorems05:17
- L10.5 Independence03:35
- L10.6 Stick-Breaking Example10:02
- L10.7 Independent Normals05:36
- L10.8 Bayes Rule Variations03:27
- L10.9 Mixed Bayes Rule08:33
- L10.10 Detection of a Binary Signal09:15
- L10.11 Inference of the Bias of a Coin06:00
- L11.1 Lecture Overview01:52
- L11.2 The PMF of a Function of a Discrete Random Variable06:42
- L11.3 A Linear Function of a Continuous Random Variable11:18
- L11.4 A Linear Function of a Normal Random Variable02:45
- L11.5 The PDF of a General Function09:47
- L11.6 The Monotonic Case11:07
- L11.7 The Intuition for the Monotonic Case05:28
- L11.8 A Nonmonotonic Example07:14
- L11.9 The PDF of a Function of Multiple Random Variables07:42
- S11.1 Simulation12:35
- L12.1 Lecture Overview01:29
- L12.2 The Sum of Independent Discrete Random Variables07:52
- L12.3 The Sum of Independent Continuous Random Variables06:45
- L12.4 The Sum of Independent Normal Random Variables03:10
- L12.5 Covariance05:54
- L12.6 Covariance Properties05:48
- L12.7 The Variance of the Sum of Random Variables05:36
- L12.8 The Correlation Coefficient07:03
- L12.9 Proof of Key Properties of the Correlation Coefficient03:52
- L12.10 Interpreting the Correlation Coefficient05:50
- L12.11 Correlations Matter06:22
- L13.1 Lecture Overview01:47
- L13.2 Conditional Expectation as a Random Variable04:31
- L13.3 The Law of Iterated Expectations03:58
- L13.4 Stick-Breaking Revisited03:53
- L13.5 Forecast Revisions04:38
- L13.6 The Conditional Variance05:02
- L13.7 Derivation of the Law of Total Variance04:54
- L13.8 A Simple Example06:29
- L13.9 Section Means and Variances09:04
- L13.10 Mean of the Sum of a Random Number of Random Variables06:26
- L13.11 Variance of the Sum of a Random Number of Random Variables05:10
- S13.1 Conditional Expectation Properties08:13
- L14.1 Lecture Overview02:10
- L14.2 Overview of Some Application Domains05:17
- L14.3 Types of Inference Problems05:24
- L14.4 The Bayesian Inference Framework09:48
- L14.5 Discrete Parameter, Discrete Observation06:46
- L14.6 Discrete Parameter, Continuous Observation04:35
- L14.7 Continuous Parameter, Continuous Observation03:46
- L14.8 Inferring the Unknown Bias of a Coin and the Beta Distribution07:35
- L14.9 Inferring the Unknown Bias of a Coin - Point Estimates09:30
- L14.10 Summary05:41
- S14.1 The Beta Formula10:24
- L15.1 Lecture Overview01:59
- L15.2 Recognizing Normal PDFs07:15
- L15.3 Estimating a Normal Random Variable in the Presence of Additive Noise08:18
- L15.4 The Case of Multiple Observations13:47
- L15.5 The Mean Squared Error13:02
- L15.6 Multiple Parameters; Trajectory Estimation10:32
- L15.7 Linear Normal Models05:12
- L15.8 Trajectory Estimation Illustration10:55
- L16.1 Lecture Overview01:13
- L16.2 LMS Estimation in the Absence of Observations06:48
- L16.3 LMS Estimation of One Random Variable Based on Another09:24
- L16.4 LMS Performance Evaluation04:32
- L16.5 Example: The LMS Estimate06:31
- L16.6 Example Continued: LMS Performance Evaluation05:29
- L16.7 LMS Estimation with Multiple Observations or Unknowns05:24
- L16.8 Properties of the LMS Estimation Error05:59
- L17.1 Lecture Overview01:41
- L17.2 LLMS Formulation04:58
- L17.3 Solution to the LLMS Problem05:06
- L17.4 Remarks on the LLMS Solution and on the Error Variance08:02
- L17.5 LLMS Example06:43
- L17.6 LLMS for Inferring the Parameter of a Coin11:29
- L17.7 LLMS with Multiple Observations06:54
- L17.8 The Simplest LLMS Example with Multiple Observations05:06
- L17.9 The Representation of the Data Matters in LLMS07:03
- L18.1 Lecture Overview01:57
- L18.2 The Markov Inequality10:21
- L18.3 The Chebyshev Inequality05:57
- L18.4 The Weak Law of Large Numbers07:31
- L18.5 Polling08:12
- L18.6 Convergence in Probability08:28
- L18.7 Convergence in Probability Examples08:05
- L18.8 Related Topics06:44
- S18.1 Convergence in Probability of the Sum of Two Random Variables10:13
- S18.2 Jensen's Inequality12:19
- S18.3 Hoeffding's Inequality18:28
- L19.1 Lecture Overview01:50
- L19.2 The Central Limit Theorem06:58
- L19.3 Discussion of the CLT09:00
- L19.4 Illustration of the CLT02:54
- L19.5 CLT Examples13:56
- L19.6 Normal Approximation to the Binomial11:53
- L19.7 Polling Revisited13:54
- L20.1 Lecture Overview02:46
- L20.2 Overview of the Classical Statistical Framework11:00
- L20.3 The Sample Mean and Some Terminology04:58
- L20.4 On the Mean Squared Error of an Estimator06:53
- L20.5 Confidence Intervals05:04
- L20.6 Confidence Intervals for the Estimation of the Mean04:27
- L20.7 Confidence Intervals for the Mean, When the Variance is Unknown06:13
- L20.8 Other Natural Estimators04:37
- L20.9 Maximum Likelihood Estimation06:32
- L20.10 Maximum Likelihood Estimation Examples10:20
- L21.1 Lecture Overview02:01
- L21.2 The Bernoulli Process04:21
- L21.3 Stochastic Processes06:21
- L21.4 Review of Known Properties of the Bernoulli Process02:20
- L21.5 The Fresh Start Property11:26
- L21.6 Example: The Distribution of a Busy Period04:16
- L21.7 The Time of the K-th Arrival08:12
- L21.8 Merging of Bernoulli Processes07:12
- L21.9 Splitting a Bernoulli Process05:54
- L21.10 The Poisson Approximation to the Binomial06:12
- L22.1 Lecture Overview01:31
- L22.2 Definition of the Poisson Process05:07
- L22.3 Applications of the Poisson Process03:03
- L22.4 The Poisson PMF for the Number of Arrivals08:01
- L22.5 The Mean and Variance of the Number of Arrivals03:22
- L22.6 A Simple Example03:07
- L22.7 Time of the K-th Arrival10:41
- L22.8 The Fresh Start Property and Its Implications10:33
- L22.9 Summary of Results02:34
- L22.10 An Example14:08
- L23.1 Lecture Overview01:39
- L23.2 The Sum of Independent Poisson Random Variables04:03
- L23.3 Merging Independent Poisson Processes08:22
- L23.4 Where is an Arrival of the Merged Process Coming From?05:00
- L23.5 The Time Until the First (or last) Lightbulb Burns Out11:25
- L23.6 Splitting a Poisson Process05:06
- L23.7 Random Incidence in the Poisson Process09:09
- L23.8 Random Incidence in a Non-Poisson Process04:36
- L23.9 Different Sampling Methods can Give Different Results03:59
- S23.1 Poisson Versus Normal Approximations to the Binomial08:56
- S23.2 Poisson Arrivals During an Exponential Interval09:37
- L24.1 Lecture Overview01:59
- L24.2 Introduction to Markov Processes02:09
- L24.3 Checkout Counter Example12:10
- L24.4 Discrete-Time Finite-State Markov Chains07:53
- L24.5 N-Step Transition Probabilities10:59
- L24.6 A Numerical Example - Part I09:26
- L24.7 Generic Convergence Questions05:32
- L24.8 Recurrent and Transient States05:37
- L25.1 Brief Introduction (RES.6-012 Introduction to Probability)01:40
- L25.2 Lecture Overview01:05
- L25.3 Markov Chain Review06:15
- L25.4 The Probability of a Path06:39
- L25.5 Recurrent and Transient States: Review03:26
- L25.6 Periodic States06:49
- L25.7 Steady-State Probabilities and Convergence09:13
- L25.8 A Numerical Example - Part II03:58
- L25.9 Visit Frequency Interpretation of Steady-State Probabilities05:19
- L25.10 Birth-Death Processes - Part I08:56
- L25.11 Birth-Death Processes - Part II08:57
- L26.1 Brief Introduction (RES.6-012 Introduction to Probability)01:41
- L26.2 Lecture Overview00:40
- L26.3 Review of Steady-State Behavior09:12
- L26.4 A Numerical Example - Part III10:35
- L26.5 Design of a Phone System18:30
- L26.6 Absorption Probabilities09:58
- L26.7 Expected Time to Absorption11:30
- L26.8 Mean First Passage Time08:44
- L26.9 Gambler's Ruin11:24
