is a Thomas Bayes postdoctoral Fellow of MCML at the Chair of Computer Vision and Artificial Intelligence at TU Munich.
His research interests include regular and geometric computer vision, signal and image processing, analysis, and synthesis, and deep learning interpretability.
is a Thomas Bayes Fellow of MCML and Interim Professor at the Chair of Statistical Learning and Data Science at LMU Munich.
His research advances explainable AI (XAI) by developing theoretically grounded, efficient methods for interpreting machine learning models, focusing on scalable and reliable Shapley-based explanations.
is a Thomas Bayes postdoctoral Fellow of MCML at the Chair for AI in Healthcare and Medicine at TU Munich.
His current research centers on fusing probabilistic and Bayesian approaches with deep learning to address critical challenges in the medical domain. Key areas of his work involve enhancing model reliability through meaningful uncertainty quantification and the effective utilization of prior information. Besides methodological advances, he is also interested in real-world applications where data exhibit problems like sparsity, measurement issues, and challenges with respect to privacy.
©all images: LMU | TUM
2025-12-18 - Last modified: 2025-12-18