1 code implementation • 9 Oct 2021 • Ahmed Frikha, Haokun Chen, Denis Krompaß, Thomas Runkler, Volker Tresp
In particular, we address the question: How can knowledge contained in models trained on different source domains be merged into a single model that generalizes well to unseen target domains, in the absence of source and target domain data?
no code implementations • 9 Sep 2021 • Ahmed Frikha, Denis Krompaß, Volker Tresp
Machine learning models that can generalize to unseen domains are essential when applied in real-world scenarios involving strong domain shifts.
1 code implementation • 10 Aug 2020 • Ahmed Frikha, Denis Krompaß, Volker Tresp
Although continual learning and anomaly detection have separately been well-studied in previous works, their intersection remains rather unexplored.
1 code implementation • 8 Jul 2020 • Ahmed Frikha, Denis Krompaß, Hans-Georg Köpken, Volker Tresp
Our experiments on eight datasets from the image and time-series domains show that our method leads to better results than classical OCC and few-shot classification approaches, and demonstrate the ability to learn unseen tasks from only few normal class samples.
no code implementations • 21 Dec 2015 • Cristóbal Esteban, Volker Tresp, Yinchong Yang, Stephan Baier, Denis Krompaß
By predicting future events, we also predict likely changes in the knowledge graph and thus obtain a model for the evolution of the knowledge graph as well.
no code implementations • 25 Nov 2015 • Volker Tresp, Cristóbal Esteban, Yinchong Yang, Stephan Baier, Denis Krompaß
We introduce a number of hypotheses on human memory that can be derived from the developed mathematical models.
no code implementations • 11 Aug 2015 • Denis Krompaß, Stephan Baier, Volker Tresp
Latent variable models have increasingly gained attention for the statistical modeling of knowledge graphs, showing promising results in tasks related to knowledge graph completion and cleaning.