* Die Zahl der Credits kann in Einzelfällen studiengangsspezifisch variieren. Es gilt der im Transcript of Records oder Leistungsnachweis ausgewiesene Wert.
Beschreibung der Studien-/Prüfungsleistungen:
The module examination is based on a written exam (60 minutes). In the exam, students are asked to write statistical algorithms to solve specific problems in a similar fashion as they have been performed in the homework. They may be asked to interpret R code and output, demonstrating that they have successfully learned how to program and interpret the output of packages in R. They are asked to recall the definitions of the important algorithms, such as the Gibbs sampler or the Metropolis-Hastings algorithm, the EM-algorithm and bootstrap.
Im Folgesemester: Nein
Am Semesterende: Ja
Introductory and advanced statistics course (e.g. MA0009, MA2404), R statistical software
Upon completion of the module, students
- know how discrete and continuous random variables/vectors are generated using statistical software such as R
- understand Bayesian principles, such as prior, posterior distributions
- understand the theory of MCMC algorithms from selected examples
- are able to construct MCMC algorithms to simulate from the posterior distributions and to assess convergence of MCMC simulations
- know how to use Bootstrap and Jacknife methods to estimate standard errors of estimators
- know how to apply the EM algorithm to missing data problems
- are able to program statistical algorithms in the statistical software package R
Computational statistics methods are required when analyzing complex data structures. In this course you will learn the basics of recent computational statistics methods such as Markov Chain Monte Carlo (MCMC) methods, expectation-maximization (EM) algorithm and the bootstrap. Emphasis will be given to basic theory and applications. In particular the following topics will be covered: Random variable generation: discrete, continuous, univariate, multivariate, resampling. Numerical methods for integration, root-finding and optimization. Bayesian inference: posterior distribution, hierarchical models, Markov chains, stationary and limiting distributions, Markov Chain Monte Carlo Methods (MCMC): Gibbs sampling, Metropolis-Hastings algorithm, implementation, convergence diagnostics, software for MCMC, Model adequacy and model choice. EM Algorithm: Theory, EM in exponential family, computation of standard errors. Bootstrap and Jacknife methods: empirical distribution and plug-in, bootstrap estimate of standard errors, jacknife and relationship to bootstrap, confidence intervals based on bootstrap percentiles, permutation tests and extensions.
Lehr- und Lernmethode:
The module is offered as lectures with accompanying practice sessions. In the lectures, the contents will be presented in a talk with illustrative examples, as well as through discussion with the students. The lectures should motivate the students to carry out their own analysis of the themes presented and to independently study the relevant literature. Attached to the lecture, practice sessions will be offered, in which exercise sheets and solutions will be available. In this way, students can deepen their understanding of the methods and concepts taught in the lectures and independently check their progress.
blackboard and slides
Rizzo ML. Statistical Computing with R, 2nd ed, 2019, CRC Press
Lehrveranstaltungen (Lehrform, SWS) Dozent(in):
820954919 Übungen zu Computational Statistics [MA3402] (1SWS UE, SS 2020/21)
Ankerst D, Miller G
820965689 Computational Statistics [MA3402] (2SWS VO, SS 2020/21)
Ankerst D, Miller G