Home | Research | Groups | Volker Tresp

Research Group Volker Tresp

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models

Volker Tresp

is Guest Professor for Machine Learning at the Chair of Database Systems & Data Mining at LMU Munich.

His team has a long-standing tradition in machine learning for relational structured domains. They particularly focus on (temporal) knowledge graphs and are currently investigating synergies with large language models. Driven by the interest in cognitive AI, they are increasingly exploring multimodal foundation models. The ultimate goal is to achieve a better understanding of human-level intelligence.

Team members @MCML

Link to Shuo Chen

Shuo Chen

Database Systems & Data Mining

A3 | Computational Models

Link to Zifeng Ding

Zifeng Ding

Database Systems & Data Mining

A3 | Computational Models

Link to Rajat Koner

Rajat Koner

Database Systems & Data Mining

A3 | Computational Models

Link to Hang Li

Hang Li

Database Systems & Data Mining

A3 | Computational Models

Link to Ruotong Liao

Ruotong Liao

Database Systems & Data Mining

A3 | Computational Models

Link to Tong Liu

Tong Liu

Database Systems & Data Mining

A3 | Computational Models

Link to Yize Sun

Yize Sun

Database Systems & Data Mining

A3 | Computational Models

Link to Gengyuan Zhang

Gengyuan Zhang

Database Systems & Data Mining

A3 | Computational Models

Link to Yao Zhang

Yao Zhang

Database Systems & Data Mining

A3 | Computational Models

Publications @MCML

[43]
Z. Ding, J. Wu, J. Wu, Y. Xia and V. Tresp.
Temporal Fact Reasoning over Hyper-Relational Knowledge Graphs.
Findings of the Conference on Empirical Methods in Natural Language Processing (EMNLP 2024). Miami, FL, USA, Nov 12-16, 2024. To be published. Preprint at arXiv. arXiv.
Abstract

Stemming from traditional knowledge graphs (KGs), hyper-relational KGs (HKGs) provide additional key-value pairs (i.e., qualifiers) for each KG fact that help to better restrict the fact validity. In recent years, there has been an increasing interest in studying graph reasoning over HKGs. Meanwhile, as discussed in recent works that focus on temporal KGs (TKGs), world knowledge is ever-evolving, making it important to reason over temporal facts in KGs. Previous mainstream benchmark HKGs do not explicitly specify temporal information for each HKG fact. Therefore, almost all existing HKG reasoning approaches do not devise any module specifically for temporal reasoning. To better study temporal fact reasoning over HKGs, we propose a new type of data structure named hyper-relational TKG (HTKG). Every fact in an HTKG is coupled with a timestamp explicitly indicating its time validity. We develop two new benchmark HTKG datasets, i.e., Wiki-hy and YAGO-hy, and propose an HTKG reasoning model that efficiently models hyper-relational temporal facts. To support future research on this topic, we open-source our datasets and model.

MCML Authors
Link to Zifeng Ding

Zifeng Ding

Database Systems & Data Mining

A3 | Computational Models

Link to Yan Xia

Yan Xia

Dr.

Computer Vision & Artificial Intelligence

B1 | Computer Vision

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[42]
R. Liao, M. Erler, H. Wang, G. Zhai, G. Zhang, Y. Ma and V. Tresp.
VideoINSTA: Zero-shot Long Video Understanding via Informative Spatial-Temporal Reasoning with LLMs.
Findings of the Conference on Empirical Methods in Natural Language Processing (EMNLP 2024). Miami, FL, USA, Nov 12-16, 2024. To be published. Preprint at arXiv. arXiv. GitHub.
Abstract

In the video-language domain, recent works in leveraging zero-shot Large Language Model-based reasoning for video understanding have become competitive challengers to previous end-to-end models. However, long video understanding presents unique challenges due to the complexity of reasoning over extended timespans, even for zero-shot LLM-based approaches. The challenge of information redundancy in long videos prompts the question of what specific information is essential for large language models (LLMs) and how to leverage them for complex spatial-temporal reasoning in long-form video analysis. We propose a framework VideoINSTA, i.e. INformative Spatial-TemporAl Reasoning for zero-shot long-form video understanding. VideoINSTA contributes (1) a zero-shot framework for long video understanding using LLMs; (2) an event-based temporal reasoning and content-based spatial reasoning approach for LLMs to reason over spatial-temporal information in videos; (3) a self-reflective information reasoning scheme balancing temporal factors based on information sufficiency and prediction confidence. Our model significantly improves the state-of-the-art on three long video question-answering benchmarks: EgoSchema, NextQA, and IntentQA, and the open question answering dataset ActivityNetQA.

MCML Authors
Link to Ruotong Liao

Ruotong Liao

Database Systems & Data Mining

A3 | Computational Models

Link to Guangyao Zhai

Guangyao Zhai

Computer Aided Medical Procedures & Augmented Reality

C1 | Medicine

Link to Gengyuan Zhang

Gengyuan Zhang

Database Systems & Data Mining

A3 | Computational Models

Link to Yunpu Ma

Yunpu Ma

Dr.

Artificial Intelligence & Machine Learning

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[41]
H. Zhang, J. Liu, Z. Han, S. Chen, B. He, V. Tresp, Z. Xu and J. Gu.
Visual Question Decomposition on Multimodal Large Language Models.
Findings of the Conference on Empirical Methods in Natural Language Processing (EMNLP 2024). Miami, FL, USA, Nov 12-16, 2024. To be published. Preprint at arXiv. arXiv.
Abstract

Question decomposition has emerged as an effective strategy for prompting Large Language Models (LLMs) to answer complex questions. However, while existing methods primarily focus on unimodal language models, the question decomposition capability of Multimodal Large Language Models (MLLMs) has yet to be explored. To this end, this paper explores visual question decomposition on MLLMs. Specifically, we introduce a systematic evaluation framework including a dataset and several evaluation criteria to assess the quality of the decomposed sub-questions, revealing that existing MLLMs struggle to produce high-quality sub-questions. To address this limitation, we propose a specific finetuning dataset, DecoVQA+, for enhancing the model's question decomposition capability. Aiming at enabling models to perform appropriate selective decomposition, we propose an efficient finetuning pipeline. The finetuning pipeline consists of our proposed dataset and a training objective for selective decomposition. Finetuned MLLMs demonstrate significant improvements in the quality of sub-questions and the policy of selective question decomposition. Additionally, the models also achieve higher accuracy with selective decomposition on VQA benchmark datasets.

MCML Authors
Link to Shuo Chen

Shuo Chen

Database Systems & Data Mining

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[40]
T. Decker, A. R. Bhattarai, J. Gu, V. Tresp and F. Buettner.
Provably Better Explanations with Optimized Aggregation of Feature Attributions.
41st International Conference on Machine Learning (ICML 2024). Vienna, Austria, Jul 21-27, 2024. URL.
MCML Authors
Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[39]
Y. Sun, J. Liu, Z. Wu, Z. Ding, Y. Ma, T. Seidl and V. Tresp.
SA-DQAS: Self-attention Enhanced Differentiable Quantum Architecture Search.
Workshop Differentiable Almost Everything: Differentiable Relaxations, Algorithms, Operators, and Simulators at the 41st International Conference on Machine Learning (ICML 2024). Vienna, Austria, Jul 21-27, 2024. PDF.
MCML Authors
Link to Yize Sun

Yize Sun

Database Systems & Data Mining

A3 | Computational Models

Link to Zifeng Ding

Zifeng Ding

Database Systems & Data Mining

A3 | Computational Models

Link to Yunpu Ma

Yunpu Ma

Dr.

Artificial Intelligence & Machine Learning

A3 | Computational Models

Link to Thomas Seidl

Thomas Seidl

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[38]
H. Li, C. Shen, P. Torr, V. Tresp and J. Gu.
Self-Discovering Interpretable Diffusion Latent Directions for Responsible Text-to-Image Generation.
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2024). Seattle, WA, USA, Jun 17-21, 2024. DOI. GitHub.
Abstract

Diffusion-based models have gained significant popularity for text-to-image generation due to their exceptional image-generation capabilities. A risk with these models is the potential generation of inappropriate content, such as biased or harmful images. However, the underlying reasons for generating such undesired content from the perspective of the diffusion model's internal representation remain unclear. Previous work interprets vectors in an interpretable latent space of diffusion models as semantic concepts. However, existing approaches cannot discover directions for arbitrary concepts, such as those related to inappropriate concepts. In this work, we propose a novel self-supervised approach to find interpretable latent directions for a given concept. With the discovered vectors, we further propose a simple approach to mitigate inappropriate generation. Extensive experiments have been conducted to verify the effectiveness of our mitigation approach, namely, for fair generation, safe generation, and responsible text-enhancing generation.

MCML Authors
Link to Hang Li

Hang Li

Database Systems & Data Mining

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[37]
Z. Ding, H. Cai, J. Wu, Y. Ma, R. Liao, B. Xiong and V. Tresp.
zrLLM: Zero-Shot Relational Learning on Temporal Knowledge Graphs with Large Language Models.
Annual Conference of the North American Chapter of the Association for Computational Linguistics (NAACL 2024). Mexico City, Mexico, Jun 16-21, 2024. URL.
Abstract

Modeling evolving knowledge over temporal knowledge graphs (TKGs) has become a heated topic. Various methods have been proposed to forecast links on TKGs. Most of them are embedding-based, where hidden representations are learned to represent knowledge graph (KG) entities and relations based on the observed graph contexts. Although these methods show strong performance on traditional TKG forecasting (TKGF) benchmarks, they face a strong challenge in modeling the unseen zero-shot relations that have no prior graph context. In this paper, we try to mitigate this problem as follows. We first input the text descriptions of KG relations into large language models (LLMs) for generating relation representations, and then introduce them into embedding-based TKGF methods. LLM-empowered representations can capture the semantic information in the relation descriptions. This makes the relations, whether seen or unseen, with similar semantic meanings stay close in the embedding space, enabling TKGF models to recognize zero-shot relations even without any observed graph context. Experimental results show that our approach helps TKGF models to achieve much better performance in forecasting the facts with previously unseen relations, while still maintaining their ability in link forecasting regarding seen relations.

MCML Authors
Link to Zifeng Ding

Zifeng Ding

Database Systems & Data Mining

A3 | Computational Models

Link to Yunpu Ma

Yunpu Ma

Dr.

Artificial Intelligence & Machine Learning

A3 | Computational Models

Link to Ruotong Liao

Ruotong Liao

Database Systems & Data Mining

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[36]
R. Liao, X. Jia, Y. Li, Y. Ma and V. Tresp.
GenTKG: Generative Forecasting on Temporal Knowledge Graph.
Annual Conference of the North American Chapter of the Association for Computational Linguistics (NAACL 2024). Mexico City, Mexico, Jun 16-21, 2024. URL. GitHub.
Abstract

The rapid advancements in large language models (LLMs) have ignited interest in the temporal knowledge graph (tKG) domain, where conventional embedding-based and rule-based methods dominate. The question remains open of whether pre-trained LLMs can understand structured temporal relational data and replace them as the foundation model for temporal relational forecasting. Therefore, we bring temporal knowledge forecasting into the generative setting. However, challenges occur in the huge chasms between complex temporal graph data structure and sequential natural expressions LLMs can handle, and between the enormous data sizes of tKGs and heavy computation costs of finetuning LLMs. To address these challenges, we propose a novel retrieval-augmented generation framework named GenTKG combining a temporal logical rule-based retrieval strategy and few-shot parameter-efficient instruction tuning to solve the above challenges, respectively. Extensive experiments have shown that GenTKG outperforms conventional methods of temporal relational forecasting with low computation resources using extremely limited training data as few as 16 samples. GenTKG also highlights remarkable cross-domain generalizability with outperforming performance on unseen datasets without re-training, and in-domain generalizability regardless of time split in the same dataset. Our work reveals the huge potential of LLMs in the tKG domain and opens a new frontier for generative forecasting on tKGs.

MCML Authors
Link to Ruotong Liao

Ruotong Liao

Database Systems & Data Mining

A3 | Computational Models

Link to Yunpu Ma

Yunpu Ma

Dr.

Artificial Intelligence & Machine Learning

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[35]
S. Chen, Z. Han, B. He, M. Buckley, P. Torr, V. Tresp and J. Gu.
Understanding and Improving In-Context Learning on Vision-language Models.
Workshop on Mathematical and Empirical Understanding of Foundation Models (ME-FoMo 2024) at the 12th International Conference on Learning Representations (ICLR 2024). Vienna, Austria, May 07-11, 2024. URL.
Abstract

Recently, in-context learning (ICL) on large language models (LLMs) has received great attention, and this technique can also be applied to vision-language models (VLMs) built upon LLMs. These VLMs can respond to queries by conditioning responses on a series of multimodal demonstrations, which comprise images, queries, and answers. Though ICL has been extensively studied on LLMs, its research on VLMs remains limited. The inclusion of additional visual information in the demonstrations motivates the following research questions: which of the two modalities in the demonstration is more significant? How can we select effective multimodal demonstrations to enhance ICL performance? This study investigates the significance of both visual and language information. Our findings indicate that ICL in VLMs is predominantly driven by the textual information in the demonstrations whereas the visual information in the demonstrations barely affects the ICL performance. Subsequently, we provide an understanding of the findings by analyzing the model information flow and comparing model inner states given different ICL settings. Motivated by our analysis, we propose a simple yet effective approach, termed Mixed Modality In-Context Example Selection (MMICES), which considers both visual and language modalities when selecting demonstrations and shows better ICL performance. Extensive experiments are conducted to support our findings, understanding, and improvement of the ICL performance of VLMs.

MCML Authors
Link to Shuo Chen

Shuo Chen

Database Systems & Data Mining

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[34]
S. Chen, Z. Han, B. He, Z. Ding, W. Yu, P. Torr, V. Tresp and J. Gu.
Red Teaming GPT-4V: Are GPT-4V Safe Against Uni/Multi-Modal Jailbreak Attacks?.
Workshop on Secure and Trustworthy Large Language Models (SeT LLM 2024) at the 12th International Conference on Learning Representations (ICLR 2024). Vienna, Austria, May 07-11, 2024. URL.
Abstract

Various jailbreak attacks have been proposed to red-team Large Language Models (LLMs) and revealed the vulnerable safeguards of LLMs. Besides, some methods are not limited to the textual modality and extend the jailbreak attack to Multimodal Large Language Models (MLLMs) by perturbing the visual input. However, the absence of a universal evaluation benchmark complicates the performance reproduction and fair comparison. Besides, there is a lack of comprehensive evaluation of closed-source state-of-the-art (SOTA) models, especially MLLMs, such as GPT-4V. To address these issues, this work first builds a comprehensive jailbreak evaluation dataset with 1445 harmful questions covering 11 different safety policies. Based on this dataset, extensive red-teaming experiments are conducted on 11 different LLMs and MLLMs, including both SOTA proprietary models and open-source models. We then conduct a deep analysis of the evaluated results and find that (1) GPT4 and GPT-4V demonstrate better robustness against jailbreak attacks compared to open-source LLMs and MLLMs. (2) Llama2 and Qwen-VL-Chat are more robust compared to other open-source models. (3) The transferability of visual jailbreak methods is relatively limited compared to textual jailbreak methods.

MCML Authors
Link to Shuo Chen

Shuo Chen

Database Systems & Data Mining

A3 | Computational Models

Link to Zifeng Ding

Zifeng Ding

Database Systems & Data Mining

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[33]
H. Chen, Y. Zhang, D. Krompass, J. Gu and V. Tresp.
FedDAT: An Approach for Foundation Model Finetuning in Multi-Modal Heterogeneous Federated Learning.
38th Conference on Artificial Intelligence (AAAI 2024). Vancouver, Canada, Feb 20-27, 2024. DOI.
Abstract

Recently, foundation models have exhibited remarkable advancements in multi-modal learning. These models, equipped with millions (or billions) of parameters, typically require a substantial amount of data for finetuning. However, collecting and centralizing training data from diverse sectors becomes challenging due to distinct privacy regulations. Federated Learning (FL) emerges as a promising solution, enabling multiple clients to collaboratively train neural networks without centralizing their local data. To alleviate client computation burdens and communication overheads, previous works have adapted Parameter-efficient Finetuning (PEFT) methods for FL. Hereby, only a small fraction of the model parameters are optimized and communicated during federated communications. Nevertheless, most previous works have focused on a single modality and neglected one common phenomenon, i.e., the presence of data heterogeneity across the clients. Therefore, in this work, we propose a finetuning framework tailored to heterogeneous multi-modal FL, called Federated Dual-Aadapter Teacher (FedDAT). Specifically, our approach leverages a Dual-Adapter Teacher (DAT) to address data heterogeneity by regularizing the client local updates and applying Mutual Knowledge Distillation (MKD) for an efficient knowledge transfer. FedDAT is the first approach that enables an efficient distributed finetuning of foundation models for a variety of heterogeneous Vision-Language tasks. To demonstrate its effectiveness, we conduct extensive experiments on four multi-modality FL benchmarks with different types of data heterogeneity, where FedDAT substantially outperforms the existing centralized PEFT methods adapted for FL.

MCML Authors
Link to Yao Zhang

Yao Zhang

Database Systems & Data Mining

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[32]
M. Bernhard, R. Amoroso, Y. Kindermann, M. Schubert, L. Baraldi, R. Cucchiara and V. Tresp.
What’s Outside the Intersection? Fine-grained Error Analysis for Semantic Segmentation Beyond IoU.
IEEE/CVF Winter Conference on Applications of Computer Vision (WACV 2024). Waikoloa, Hawaii, Jan 04-08, 2024. DOI. GitHub.
Abstract

Semantic segmentation represents a fundamental task in computer vision with various application areas such as autonomous driving, medical imaging, or remote sensing. For evaluating and comparing semantic segmentation models, the mean intersection over union (mIoU) is currently the gold standard. However, while mIoU serves as a valuable benchmark, it does not offer insights into the types of errors incurred by a model. Moreover, different types of errors may have different impacts on downstream applications. To address this issue, we propose an intuitive method for the systematic categorization of errors, thereby enabling a fine-grained analysis of semantic segmentation models. Since we assign each erroneous pixel to precisely one error type, our method seamlessly extends the popular IoU-based evaluation by shedding more light on the false positive and false negative predictions. Our approach is model- and dataset-agnostic, as it does not rely on additional information besides the predicted and ground-truth segmentation masks. In our experiments, we demonstrate that our method accurately assesses model strengths and weaknesses on a quantitative basis, thus reducing the dependence on time-consuming qualitative model inspection. We analyze a variety of state-of-the-art semantic segmentation models, revealing systematic differences across various architectural paradigms. Exploiting the gained insights, we showcase that combining two models with complementary strengths in a straightforward way is sufficient to consistently improve mIoU, even for models setting the current state of the art on ADE20K.

MCML Authors
Link to Maximilian Bernhard

Maximilian Bernhard

Database Systems & Data Mining

A3 | Computational Models

Link to Matthias Schubert

Matthias Schubert

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[31]
U. Sahin, H. Li, Q. Khan, D. Cremers and V. Tresp.
Enhancing Multimodal Compositional Reasoning of Visual Language Models With Generative Negative Mining.
IEEE/CVF Winter Conference on Applications of Computer Vision (WACV 2024). Waikoloa, Hawaii, Jan 04-08, 2024. DOI. GitHub.
MCML Authors
Link to Hang Li

Hang Li

Database Systems & Data Mining

A3 | Computational Models

Link to Qadeer Khan

Qadeer Khan

Computer Vision & Artificial Intelligence

B1 | Computer Vision

Link to Daniel Cremers

Daniel Cremers

Prof. Dr.

Computer Vision & Artificial Intelligence

B1 | Computer Vision

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[30]
G. Zhang, Y. Zhang, K. Zhang and V. Tresp.
Can Vision-Language Models be a Good Guesser? Exploring VLMs for Times and Location Reasoning.
IEEE/CVF Winter Conference on Applications of Computer Vision (WACV 2024). Waikoloa, Hawaii, Jan 04-08, 2024. DOI.
MCML Authors
Link to Gengyuan Zhang

Gengyuan Zhang

Database Systems & Data Mining

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[29]
S. Chen, J. Gu, Z. Han, Y. Ma, P. Torr and V. Tresp.
Benchmarking Robustness of Adaptation Methods on Pre-trained Vision-Language Models.
37th Conference on Neural Information Processing Systems (NeurIPS 2023). New Orleans, LA, USA, Dec 10-16, 2023. URL. GitHub.
MCML Authors
Link to Shuo Chen

Shuo Chen

Database Systems & Data Mining

A3 | Computational Models

Link to Yunpu Ma

Yunpu Ma

Dr.

Artificial Intelligence & Machine Learning

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[28]
R. Liao, X. Jia, Y. Ma and V. Tresp.
GenTKG: Generative Forecasting on Temporal Knowledge Graph.
Workshop New Frontiers in Graph Learning (GLFrontiers 2023) at the 37th Conference on Neural Information Processing Systems (NeurIPS 2023). New Orleans, LA, USA, Dec 10-16, 2023. URL.
MCML Authors
Link to Ruotong Liao

Ruotong Liao

Database Systems & Data Mining

A3 | Computational Models

Link to Yunpu Ma

Yunpu Ma

Dr.

Artificial Intelligence & Machine Learning

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[27]
H. Chen, A. Frikha, D. Krompass, J. Gu and V. Tresp.
FRAug: Tackling Federated Learning with Non-IID Features via Representation Augmentation.
IEEE/CVF International Conference on Computer Vision (ICCV 2023). Paris, France, Oct 02-06, 2023. DOI.
MCML Authors
Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[26]
H. Li, J. Gu, R. Koner, S. Sharifzadeh and V. Tresp.
Do DALL-E and Flamingo Understand Each Other?.
IEEE/CVF International Conference on Computer Vision (ICCV 2023). Paris, France, Oct 02-06, 2023. DOI.
MCML Authors
Link to Hang Li

Hang Li

Database Systems & Data Mining

A3 | Computational Models

Link to Rajat Koner

Rajat Koner

Database Systems & Data Mining

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[25]
G. Zhang, J. Ren, J. Gu and V. Tresp.
Multi-event Video-Text Retrieval.
IEEE/CVF International Conference on Computer Vision (ICCV 2023). Paris, France, Oct 02-06, 2023. DOI. GitHub.
MCML Authors
Link to Gengyuan Zhang

Gengyuan Zhang

Database Systems & Data Mining

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[24]
Y. Shen, R. Liao, Z. Han, Y. Ma and V. Tresp.
GraphextQA: A Benchmark for Evaluating Graph-Enhanced Large Language Models.
Preprint at arXiv (Oct. 2023). arXiv.
MCML Authors
Link to Ruotong Liao

Ruotong Liao

Database Systems & Data Mining

A3 | Computational Models

Link to Yunpu Ma

Yunpu Ma

Dr.

Artificial Intelligence & Machine Learning

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[23]
A. Giovagnoli, Y. Ma, M. Schubert and V. Tresp.
QNEAT: Natural Evolution of Variational Quantum Circuit Architecture.
Genetic and Evolutionary Computation Conference (GECCO 2023). Lisbon, Portugal, Jul 15-19, 2023. DOI.
Abstract

Quantum Machine Learning (QML) is a recent and rapidly evolving field where the theoretical framework and logic of quantum mechanics is employed to solve machine learning tasks. A variety of techniques that have a different level of quantum-classical hybridization has been presented. Here we focus on variational quantum circuits (VQC), which emerged as the most promising candidates for the quantum counterpart of neural networks in the noisy intermediate-scale quantum (NISQ) era. Although showing promising results, VQCs can be hard to train because of different issues e.g. barren plateau, periodicity of the weights or choice of the architecture. In this paper we focus on this last problem and in order to address it we propose a gradient free algorithm inspired by natural evolution to optimise both the weights and the architecture of the VQC. In particular, we present a version of the well known neuroevolution of augmenting topologies (NEAT) algorithm adapted to the case of quantum variational circuits. We test the algorithm with different benchmark problems of classical fields of machine learning i.e. reinforcement learning and optimization.

MCML Authors
Link to Yunpu Ma

Yunpu Ma

Dr.

Artificial Intelligence & Machine Learning

A3 | Computational Models

Link to Matthias Schubert

Matthias Schubert

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[22]
Z. Han, R. Liao, J. Gu, Y. Zhang, Z. Ding, Y. Gu, H. Köppl, H. Schütze and V. Tresp.
ECOLA: Enhanced Temporal Knowledge Embeddings with Contextualized Language Representations.
Findings of the 61th Annual Meeting of the Association for Computational Linguistics (ACL 2023). Toronto, Canada, Jul 09-14, 2023. DOI.
Abstract

Since conventional knowledge embedding models cannot take full advantage of the abundant textual information, there have been extensive research efforts in enhancing knowledge embedding using texts. However, existing enhancement approaches cannot apply to temporal knowledge graphs (tKGs), which contain time-dependent event knowledge with complex temporal dynamics. Specifically, existing enhancement approaches often assume knowledge embedding is time-independent. In contrast, the entity embedding in tKG models usually evolves, which poses the challenge of aligning temporally relevant texts with entities. To this end, we propose to study enhancing temporal knowledge embedding with textual data in this paper. As an approach to this task, we propose Enhanced Temporal Knowledge Embeddings with Contextualized Language Representations (ECOLA), which takes the temporal aspect into account and injects textual information into temporal knowledge embedding. To evaluate ECOLA, we introduce three new datasets for training and evaluating ECOLA. Extensive experiments show that ECOLA significantly enhances temporal KG embedding models with up to 287% relative improvements regarding Hits@1 on the link prediction task.

MCML Authors
Link to Ruotong Liao

Ruotong Liao

Database Systems & Data Mining

A3 | Computational Models

Link to Yao Zhang

Yao Zhang

Database Systems & Data Mining

A3 | Computational Models

Link to Zifeng Ding

Zifeng Ding

Database Systems & Data Mining

A3 | Computational Models

Link to Hinrich Schütze

Hinrich Schütze

Prof. Dr.

Statistical NLP and Deep Learning

B2 | Natural Language Processing

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[21]
T. Hannan, R. Koner, M. Bernhard, S. Shit, B. Menze, V. Tresp, M. Schubert and T. Seidl.
GRAtt-VIS: Gated Residual Attention for Auto Rectifying Video Instance Segmentation.
Preprint at arXiv (May. 2023). arXiv.
MCML Authors
Link to Tanveer Hannan

Tanveer Hannan

Database Systems & Data Mining

A3 | Computational Models

Link to Rajat Koner

Rajat Koner

Database Systems & Data Mining

A3 | Computational Models

Link to Maximilian Bernhard

Maximilian Bernhard

Database Systems & Data Mining

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models

Link to Matthias Schubert

Matthias Schubert

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models

Link to Thomas Seidl

Thomas Seidl

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[20]
R. Koner, T. Hannan, S. Shit, S. Sharifzadeh, M. Schubert, T. Seidl and V. Tresp.
InstanceFormer: An Online Video Instance Segmentation Framework.
37th Conference on Artificial Intelligence (AAAI 2023). Washington, DC, USA, Feb 07-14, 2023. DOI. GitHub.
MCML Authors
Link to Rajat Koner

Rajat Koner

Database Systems & Data Mining

A3 | Computational Models

Link to Tanveer Hannan

Tanveer Hannan

Database Systems & Data Mining

A3 | Computational Models

Link to Matthias Schubert

Matthias Schubert

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models

Link to Thomas Seidl

Thomas Seidl

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[19]
M. Ali, M. Berrendorf, C. T. Hoyt, L. Vermue, M. Galkin, S. Sharifzadeh, A. Fischer, V. Tresp and J. Lehmann.
Bringing Light Into the Dark: A Large-scale Evaluation of Knowledge Graph Embedding Models under a Unified Framework.
IEEE Transactions on Pattern Analysis and Machine Intelligence 44.12 (Dec. 2022). DOI. GitHub.
Abstract

The heterogeneity in recently published knowledge graph embedding models’ implementations, training, and evaluation has made fair and thorough comparisons difficult. To assess the reproducibility of previously published results, we re-implemented and evaluated 21 models in the PyKEEN software package. In this paper, we outline which results could be reproduced with their reported hyper-parameters, which could only be reproduced with alternate hyper-parameters, and which could not be reproduced at all, as well as provide insight as to why this might be the case. We then performed a large-scale benchmarking on four datasets with several thousands of experiments and 24,804 GPU hours of computation time. We present insights gained as to best practices, best configurations for each model, and where improvements could be made over previously published best configurations. Our results highlight that the combination of model architecture, training approach, loss function, and the explicit modeling of inverse relations is crucial for a model’s performance and is not only determined by its architecture. We provide evidence that several architectures can obtain results competitive to the state of the art when configured carefully.

MCML Authors
Link to Max Berrendorf

Max Berrendorf

Dr.

* Former member

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[18]
S. Shit, R. Koner, B. Wittmann, J. Paetzold, I. Ezhov, H. Li, J. Pan, S. Sharifzadeh, G. Kaissis, V. Tresp and B. Menze.
Relationformer: A Unified Framework for Image-to-Graph Generation.
17th European Conference on Computer Vision (ECCV 2022). Tel Aviv, Israel, Oct 23-27, 2022. DOI. GitHub.
Abstract

A comprehensive representation of an image requires understanding objects and their mutual relationship, especially in image-to-graph generation, e.g., road network extraction, blood-vessel network extraction, or scene graph generation. Traditionally, image-to-graph generation is addressed with a two-stage approach consisting of object detection followed by a separate relation prediction, which prevents simultaneous object-relation interaction. This work proposes a unified one-stage transformer-based framework, namely Relationformer that jointly predicts objects and their relations. We leverage direct set-based object prediction and incorporate the interaction among the objects to learn an object-relation representation jointly. In addition to existing [obj]-tokens, we propose a novel learnable token, namely [rln]-token. Together with [obj]-tokens, [rln]-token exploits local and global semantic reasoning in an image through a series of mutual associations. In combination with the pair-wise [obj]-token, the [rln]-token contributes to a computationally efficient relation prediction. We achieve state-of-the-art performance on multiple, diverse and multi-domain datasets that demonstrate our approach’s effectiveness and generalizability.

MCML Authors
Link to Rajat Koner

Rajat Koner

Database Systems & Data Mining

A3 | Computational Models

Link to Georgios Kaissis

Georgios Kaissis

Dr.

Privacy-Preserving and Trustworthy AI

A1 | Statistical Foundations & Explainability

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[17]
Z. Ding, R. Qi, Z. Li, B. He, J. Wu, Y. Ma, Z. Meng, Z. Han and V. Tresp.
Forecasting Question Answering over Temporal Knowledge Graphs.
Preprint at arXiv (Aug. 2022). arXiv.
Abstract

Question answering over temporal knowledge graphs (TKGQA) has recently found increasing interest. TKGQA requires temporal reasoning techniques to extract the relevant information from temporal knowledge bases. The only existing TKGQA dataset, i.e., CronQuestions, consists of temporal questions based on the facts from a fixed time period, where a temporal knowledge graph (TKG) spanning the same period can be fully used for answer inference, allowing the TKGQA models to use even the future knowledge to answer the questions based on the past facts. In real-world scenarios, however, it is also common that given the knowledge until now, we wish the TKGQA systems to answer the questions asking about the future. As humans constantly seek plans for the future, building TKGQA systems for answering such forecasting questions is important. Nevertheless, this has still been unexplored in previous research. In this paper, we propose a novel task: forecasting question answering over temporal knowledge graphs. We also propose a large-scale TKGQA benchmark dataset, i.e., ForecastTKGQuestions, for this task. It includes three types of questions, i.e., entity prediction, yes-no, and fact reasoning questions. For every forecasting question in our dataset, QA models can only have access to the TKG information before the timestamp annotated in the given question for answer inference. We find that the state-of-the-art TKGQA methods perform poorly on forecasting questions, and they are unable to answer yes-no questions and fact reasoning questions. To this end, we propose ForecastTKGQA, a TKGQA model that employs a TKG forecasting module for future inference, to answer all three types of questions. Experimental results show that ForecastTKGQA outperforms recent TKGQA methods on the entity prediction questions, and it also shows great effectiveness in answering the other two types of questions.

MCML Authors
Link to Zifeng Ding

Zifeng Ding

Database Systems & Data Mining

A3 | Computational Models

Link to Zongyue Li

Zongyue Li

Database Systems & Data Mining

A3 | Computational Models

Link to Yunpu Ma

Yunpu Ma

Dr.

Artificial Intelligence & Machine Learning

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[16]
M. Ali, M. Berrendorf, M. Galkin, V. Thost, T. Ma, V. Tresp and J. Lehmann.
Improving Inductive Link Prediction Using Hyper-Relational Facts (Extended Abstract).
Best paper track at the 31st International Joint Conference on Artificial Intelligence and the 25th European Conference on Artificial Intelligence (IJCAI-ECAI 2022). Vienna, Austria, Jul 23-29, 2022. DOI.
MCML Authors
Link to Max Berrendorf

Max Berrendorf

Dr.

* Former member

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[15]
L. Hang, Q. Khan, V. Tresp and D. Cremers.
Biologically Inspired Neural Path Finding.
15th International Conference on Brain Informatics (BI 2022). Padova, Italy, Jul 15-15, 2022. DOI.
MCML Authors
Link to Qadeer Khan

Qadeer Khan

Computer Vision & Artificial Intelligence

B1 | Computer Vision

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models

Link to Daniel Cremers

Daniel Cremers

Prof. Dr.

Computer Vision & Artificial Intelligence

B1 | Computer Vision


[14]
C. T. Hoyt, M. Berrendorf, M. Gaklin, V. Tresp and B. M. Gyori.
A Unified Framework for Rank-based Evaluation Metrics for Link Prediction in Knowledge Graphs.
Workshop on Graph Learning Benchmarks (GLB 2022) at the International World Wide Web Conference (WWW 2022). Virtual, Apr 22-29, 2022. arXiv.
MCML Authors
Link to Max Berrendorf

Max Berrendorf

Dr.

* Former member

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[13]
Y. Liu, Y. Ma, M. Hildebrandt, M. Joblin and V. Tresp.
TLogic: Temporal logical rules for explainable link forecasting on temporal knowledge graphs.
36th Conference on Artificial Intelligence (AAAI 2022). Virtual, Feb 22-Mar 01, 2022. DOI.
MCML Authors
Link to Yunpu Ma

Yunpu Ma

Dr.

Artificial Intelligence & Machine Learning

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[12]
S. Sharifzadeh, S. M. Baharlou, M. Schmitt, H. Schütze and V. Tresp.
Improving Scene Graph Classification by Exploiting Knowledge from Texts.
36th Conference on Artificial Intelligence (AAAI 2022). Virtual, Feb 22-Mar 01, 2022. DOI.
MCML Authors
Link to Hinrich Schütze

Hinrich Schütze

Prof. Dr.

Statistical NLP and Deep Learning

B2 | Natural Language Processing

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[11]
M. Ali, M. Berrendorf, M. Galkin, V. Thost, T. Ma, V. Tresp and J. Lehmann.
Improving Inductive Link Prediction Using Hyper-Relational Facts.
20th International Semantic Web Conference (ISWC 2021). Virtual, Oct 24-28, 2021. DOI. GitHub.
Abstract

For many years, link prediction on knowledge graphs (KGs) has been a purely transductive task, not allowing for reasoning on unseen entities. Recently, increasing efforts are put into exploring semi- and fully inductive scenarios, enabling inference over unseen and emerging entities. Still, all these approaches only consider triple-based KGs, whereas their richer counterparts, hyper-relational KGs (e.g., Wikidata), have not yet been properly studied. In this work, we classify different inductive settings and study the benefits of employing hyper-relational KGs on a wide range of semi- and fully inductive link prediction tasks powered by recent advancements in graph neural networks. Our experiments on a novel set of benchmarks show that qualifiers over typed edges can lead to performance improvements of 6% of absolute gains (for the Hits@10 metric) compared to triple-only baselines.

MCML Authors
Link to Max Berrendorf

Max Berrendorf

Dr.

* Former member

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[10]
Y. Ma and V. Tresp.
Causal Inference under Networked Interference and Intervention Policy Enhancement.
24th International Conference on Artificial Intelligence and Statistics (AISTATS 2021). Virtual, Apr 13-15, 2021. URL.
MCML Authors
Link to Yunpu Ma

Yunpu Ma

Dr.

Artificial Intelligence & Machine Learning

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[9]
M. Berrendorf, E. Faerman and V. Tresp.
Active Learning for Entity Alignment.
43rd European Conference on Information Retrieval (ECIR 2021). Virtual, Mar 28-Apr 01, 2021. DOI. GitHub.
Abstract

In this work, we propose a novel framework for labeling entity alignments in knowledge graph datasets. Different strategies to select informative instances for the human labeler build the core of our framework. We illustrate how the labeling of entity alignments is different from assigning class labels to single instances and how these differences affect the labeling efficiency. Based on these considerations, we propose and evaluate different active and passive learning strategies. One of our main findings is that passive learning approaches, which can be efficiently precomputed, and deployed more easily, achieve performance comparable to the active learning strategies.

MCML Authors
Link to Max Berrendorf

Max Berrendorf

Dr.

* Former member

A3 | Computational Models

Link to Evgeny Faerman

Evgeny Faerman

Dr.

* Former member

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[8]
M. Ali, M. Berrendorf, C. T. Hoyt, L. Vermue, S. Sharifzadeh, V. Tresp and J. Lehmann.
PyKEEN 1.0: A Python Library for Training and Evaluating Knowledge Graph Embeddings.
Journal of Machine Learning Research 22.82 (Mar. 2021). PDF.
Abstract

Recently, knowledge graph embeddings (KGEs) have received significant attention, and several software libraries have been developed for training and evaluation. While each of them addresses specific needs, we report on a community effort to a re-design and re-implementation of PyKEEN, one of the early KGE libraries. PyKEEN 1.0 enables users to compose knowledge graph embedding models based on a wide range of interaction models, training approaches, loss functions, and permits the explicit modeling of inverse relations. It allows users to measure each component’s influence individually on the model’s performance. Besides, an automatic memory optimization has been realized in order to optimally exploit the provided hardware. Through the integration of Optuna, extensive hyper-parameter optimization (HPO) functionalities are provided.

MCML Authors
Link to Max Berrendorf

Max Berrendorf

Dr.

* Former member

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[7]
S. Sharifzadeh, S. M. Baharlou and V. Tresp.
Classification by Attention: Scene Graph Classification with Prior Knowledge.
35th Conference on Artificial Intelligence (AAAI 2021). Virtual, Feb 02-09, 2021. DOI.
MCML Authors
Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[6]
M. Berrendorf, E. Faerman, L. Vermue and V. Tresp.
Interpretable and Fair Comparison of Link Prediction or Entity Alignment Methods with Adjusted Mean Rank.
IEEE/WIC/ACM International Joint Conference on Web Intelligence and Intelligent Agent Technology (WI-IAT 2020). Virtual, Dec 14-17, 2020. DOI.
Abstract

In this work, we take a closer look at the evaluation of two families of methods for enriching information from knowledge graphs: Link Prediction and Entity Alignment. In the current experimental setting, multiple different scores are employed to assess different aspects of model performance. We analyze the informativeness of these evaluation measures and identify several shortcomings. In particular, we demonstrate that all existing scores can hardly be used to compare results across different datasets. Moreover, we demonstrate that varying size of the test size automatically has impact on the performance of the same model based on commonly used metrics for the Entity Alignment task. We show that this leads to various problems in the interpretation of results, which may support misleading conclusions. Therefore, we propose adjustments to the evaluation and demonstrate empirically how this supports a fair, comparable, and interpretable assessment of model performance.

MCML Authors
Link to Max Berrendorf

Max Berrendorf

Dr.

* Former member

A3 | Computational Models

Link to Evgeny Faerman

Evgeny Faerman

Dr.

* Former member

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[5]
Y. Ma and V. Tresp.
A Variational Quantum Circuit Model for Knowledge Graph Embeddings.
1st Workshop on Quantum Tensor Networks in Machine Learning (QTNML 2020) at the 34th Conference on Neural Information Processing Systems (NeurIPS 2020). Virtual, Dec 06-12, 2020. PDF.
MCML Authors
Link to Yunpu Ma

Yunpu Ma

Dr.

Artificial Intelligence & Machine Learning

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[4]
Y. Ma, Z. Han and V. Tresp.
Learning with Temporal Knowledge Graphs.
CIKM 2020 Workshops (CIKMW 2020) co-located with the 29th ACM International Conference on Information and Knowledge Management (CIKM 2020). Galway, Ireland, Oct 19-23, 2020. Invited talk. PDF.
MCML Authors
Link to Yunpu Ma

Yunpu Ma

Dr.

Artificial Intelligence & Machine Learning

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[3]
M. Berrendorf, E. Faerman and V. Tresp.
Active Learning for Entity Alignment.
5th International Workshop on Deep Learning for Graphs (DL4G@WWW2020) at the ACM Web Conference 2020 (WWW 2020). Taipeh, Taiwan, Apr 21, 2020. arXiv.
MCML Authors
Link to Max Berrendorf

Max Berrendorf

Dr.

* Former member

A3 | Computational Models

Link to Evgeny Faerman

Evgeny Faerman

Dr.

* Former member

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[2]
M. Berrendorf, E. Faerman, L. Vermue and V. Tresp.
Interpretable and Fair Comparison of Link Prediction or Entity Alignment Methods with Adjusted Mean Rank (Extended Abstract).
5th International Workshop on Deep Learning for Graphs (DL4G@WWW2020) at the ACM Web Conference 2020 (WWW 2020). Taipeh, Taiwan, Apr 21, 2020. Full papaer at WI-AT 2020. DOI.
MCML Authors
Link to Max Berrendorf

Max Berrendorf

Dr.

* Former member

A3 | Computational Models

Link to Evgeny Faerman

Evgeny Faerman

Dr.

* Former member

A3 | Computational Models

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models


[1]
M. Berrendorf, E. Faerman, V. Melnychuk, V. Tresp and T. Seidl.
Knowledge Graph Entity Alignment with Graph Convolutional Networks: Lessons Learned.
42nd European Conference on Information Retrieval (ECIR 2020). Virtual, Apr 14-17, 2020. DOI. GitHub.
Abstract

In this work, we focus on the problem of entity alignment in Knowledge Graphs (KG) and we report on our experiences when applying a Graph Convolutional Network (GCN) based model for this task. Variants of GCN are used in multiple state-of-the-art approaches and therefore it is important to understand the specifics and limitations of GCN-based models. Despite serious efforts, we were not able to fully reproduce the results from the original paper and after a thorough audit of the code provided by authors, we concluded, that their implementation is different from the architecture described in the paper. In addition, several tricks are required to make the model work and some of them are not very intuitive.We provide an extensive ablation study to quantify the effects these tricks and changes of architecture have on final performance. Furthermore, we examine current evaluation approaches and systematize available benchmark datasets.We believe that people interested in KG matching might profit from our work, as well as novices entering the field.

MCML Authors
Link to Max Berrendorf

Max Berrendorf

Dr.

* Former member

A3 | Computational Models

Link to Evgeny Faerman

Evgeny Faerman

Dr.

* Former member

A3 | Computational Models

Link to Valentyn Melnychuk

Valentyn Melnychuk

Artificial Intelligence in Management

C4 | Computational Social Sciences

Link to Volker Tresp

Volker Tresp

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models

Link to Thomas Seidl

Thomas Seidl

Prof. Dr.

Database Systems & Data Mining

A3 | Computational Models