Computational Statistics
Module MA3402
This module handbook serves to describe contents, learning outcome, methods and examination type as well as linking to current dates for courses and module examination in the respective sections.
Module version of SS 2020
There are historic module descriptions of this module. A module description is valid until replaced by a newer one.
Whether the module’s courses are offered during a specific semester is listed in the section Courses, Learning and Teaching Methods and Literature below.
available module versions | |||||
---|---|---|---|---|---|
SS 2021 | WS 2020/1 | SS 2020 | SS 2019 | SS 2012 | WS 2011/2 |
Basic Information
MA3402 is a semester module in English language at Master’s level which is offered in summer semester.
This module description is valid to WS 2021/2.
Total workload | Contact hours | Credits (ECTS) |
---|---|---|
150 h | 45 h | 5 CP |
Content, Learning Outcome and Preconditions
Content
Computational statistics methods are required when analyzing complex data structures. In this course you will learn the basics of recent computational statistics methods such as Markov Chain Monte Carlo (MCMC) methods, expectation-maximization (EM) algorithm and the bootstrap. Emphasis will be given to basic theory and applications. In particular the following topics will be covered: Random variable generation: discrete, continuous, univariate, multivariate, resampling. Numerical methods for integration, root-finding and optimization. Bayesian inference: posterior distribution, hierarchical models, Markov chains, stationary and limiting distributions, Markov Chain Monte Carlo Methods (MCMC): Gibbs sampling, Metropolis-Hastings algorithm, implementation, convergence diagnostics, software for MCMC, Model adequacy and model choice. EM Algorithm: Theory, EM in exponential family, computation of standard errors. Bootstrap and Jacknife methods: empirical distribution and plug-in, bootstrap estimate of standard errors, jacknife and relationship to bootstrap, confidence intervals based on bootstrap percentiles, permutation tests and extensions.
Learning Outcome
Upon completion of the module, students
- know how discrete and continuous random variables/vectors are generated using statistical software such as R
- understand Bayesian principles, such as prior, posterior distributions
- understand the theory of MCMC algorithms from selected examples
- are able to construct MCMC algorithms to simulate from the posterior distributions and to assess convergence of MCMC simulations
- know how to use Bootstrap and Jacknife methods to estimate standard errors of estimators
- know how to apply the EM algorithm to missing data problems
- are able to program statistical algorithms in the statistical software package R
- know how discrete and continuous random variables/vectors are generated using statistical software such as R
- understand Bayesian principles, such as prior, posterior distributions
- understand the theory of MCMC algorithms from selected examples
- are able to construct MCMC algorithms to simulate from the posterior distributions and to assess convergence of MCMC simulations
- know how to use Bootstrap and Jacknife methods to estimate standard errors of estimators
- know how to apply the EM algorithm to missing data problems
- are able to program statistical algorithms in the statistical software package R
Preconditions
Introductory and advanced statistics course (e.g. MA0009, MA2404), R statistical software
Courses, Learning and Teaching Methods and Literature
Courses and Schedule
Type | SWS | Title | Lecturer(s) | Dates | Links |
---|---|---|---|---|---|
VO | 2 | Computational Statistics [MA3402] | Ankerst, D. Miller, G. |
Tue, 10:15–11:45, virtuell |
eLearning |
UE | 1 | Computational Statistics (Exercise Session) [MA3402] | Ankerst, D. Miller, G. | dates in groups |
Learning and Teaching Methods
The module is offered as lectures with accompanying practice sessions. In the lectures, the contents will be presented in a talk with illustrative examples, as well as through discussion with the students. The lectures should motivate the students to carry out their own analysis of the themes presented and to independently study the relevant literature. Attached to the lecture, practice sessions will be offered, in which exercise sheets and solutions will be available. In this way, students can deepen their understanding of the methods and concepts taught in the lectures and independently check their progress.
Media
blackboard and slides
Literature
Rizzo ML. Statistical Computing with R, 2nd ed, 2019, CRC Press
Module Exam
Description of exams and course work
The module examination is based on a written exam (60 minutes). In the exam, students are asked to write statistical algorithms to solve specific problems in a similar fashion as they have been performed in the homework. They may be asked to interpret R code and output, demonstrating that they have successfully learned how to program and interpret the output of packages in R. They are asked to recall the definitions of the important algorithms, such as the Gibbs sampler or the Metropolis-Hastings algorithm, the EM-algorithm and bootstrap.
Exam Repetition
The exam may be repeated at the end of the semester.