DTSA 5726: Introduction to Bayesian Statistics for Data Science Applications

  • Specialization: Bayesian Statistics for Data Science
  • Instructor: Brian Zaharatos
  • Prior knowledge needed: Statistics 

 View on Coursera

Learning Outcomes

  • Articulate the primary interpretations of probability theory and the role these interpretations play in Bayesian inference
  • Use Bayesian inference to solve real-world statistics and data science problems.
  • Articulate the logic of Bayesian inference and compare and contrast it with frequentist inference.
  • Utilize conjugate, improper, and objective priors to find posterior distributions.
  • Evaluate the ethical consequences of the use (or misuse) of statistical methods.

Course Content

Duration: 11h

This module introduces learners to Bayesian statistics by comparing Bayesian and frequentist methods. The introduction is motivated by an example that illustrates how different assumptions about data collection - specifically, stopping rules - can result in different conclusions when using frequentist methods. Bayesian methods, on the other hand, yield the same conclusion regardless of stopping rules. This example illuminates a key philosophical difference between frequentist and Bayesian methods.

Duration: 7h

This module introduces learners to Bayesian inference through an example using discrete data. The example demonstrates how the posterior distribution is calculated and how uncertainty is quantified in Bayesian statistics.  The module also describes methods for summarizing the posterior distribution and introduces learners to the posterior predictive distribution through use of the Monte Carlo simulation. Monte Carlo simulations will be important for advanced computational Bayesian methods.

Duration: 6h

This module introduces learners to methods for conducting Bayesian inference when the likelihood and prior distributions come from a convenient family of distributions, called conjugate families. Conjugate families are a class of prior distributions for which the posterior distribution is in the same class.  The module covers the beta-binomial, normal-normal and inverse gamma-normal conjugate families and includes examples of their application to find posterior distributions in R.

Duration: 6h

This module motivates, defines, and utilizes improper and so-called "objective" prior distributions in Bayesian statistical inference. 

Duration: 9h

In this module, learners will be introduced to Bayesian inference involving more than one unknown parameter.  Multiparameter problems are motivated with a simple example: a conjugate prior, two-parameter model involving normally distributed data. From there, we learn to solve more complex problems, including Bayesian linear regression and variance-covariance matrix estimation.    

Duration: 2h

You will complete a proctored final exam worth 35% of your final grade. You must attempt the final in order to earn a grade in the course. If you've upgraded to the for-credit version of this course, please make sure you review the additional for-credit materials in the Introductory module and anywhere else they may be found.

Note: This page is periodically updated. Course information on the Coursera platform supersedes the information on this page. Click View on Coursera button above for the most up-to-date information.