1 code implementation • 6 Apr 2024 • Anurag Singh, Siu Lun Chau, Shahine Bouabid, Krikamol Muandet
Out-of-distribution (OOD) generalisation is challenging because it involves not only learning from empirical data, but also deciding among various notions of generalisation, e. g., optimising the average-case risk, worst-case risk, or interpolations thereof.
1 code implementation • 26 Oct 2023 • Masaki Adachi, Brady Planden, David A. Howey, Michael A. Osborne, Sebastian Orbell, Natalia Ares, Krikamol Muandet, Siu Lun Chau
Like many optimizers, Bayesian optimization often falls short of gaining user trust due to opacity.
2 code implementations • 30 Aug 2023 • Kiet Q. H. Vo, Muneeb Aadil, Siu Lun Chau, Krikamol Muandet
We study the problem of agent selection in causal strategic learning under multiple decision makers and address two key challenges that come with it.
1 code implementation • 24 Jun 2022 • Simon Föll, Alina Dubatovka, Eugen Ernst, Siu Lun Chau, Martin Maritsch, Patrik Okanovic, Gudrun Thäter, Joachim M. Buhmann, Felix Wortmann, Krikamol Muandet
To address this problem, we postulate that real-world distributions are composed of latent Invariant Elementary Distributions (I. E. D) across different domains.
1 code implementation • 26 May 2022 • Robert Hu, Siu Lun Chau, Jaime Ferrando Huertas, Dino Sejdinovic
While preference modelling is becoming one of the pillars of machine learning, the problem of preference explanation remains challenging and underexplored.
no code implementations • 2 Feb 2022 • Robert Hu, Siu Lun Chau, Dino Sejdinovic, Joan Alexis Glaunès
Kernel matrix-vector multiplication (KMVM) is a foundational operation in machine learning and scientific computing.
no code implementations • 18 Oct 2021 • Siu Lun Chau, Robert Hu, Javier Gonzalez, Dino Sejdinovic
Feature attribution for kernel methods is often heuristic and not individualised for each prediction.
no code implementations • NeurIPS 2021 • Siu Lun Chau, Jean-François Ton, Javier González, Yee Whye Teh, Dino Sejdinovic
While causal models are becoming one of the mainstays of machine learning, the problem of uncertainty quantification in causal inference remains challenging.
1 code implementation • NeurIPS 2021 • Siu Lun Chau, Shahine Bouabid, Dino Sejdinovic
Yet, when LR samples are modeled as aggregate conditional means of HR samples with respect to a mediating variable that is globally observed, the recovery of the underlying fine-grained field can be framed as taking an "inverse" of the conditional expectation, namely a deconditioning problem.
no code implementations • 23 Aug 2020 • Xingyue Pu, Siu Lun Chau, Xiaowen Dong, Dino Sejdinovic
In this paper, we propose a novel graph learning framework that incorporates the node-side and observation-side information, and in particular the covariates that help to explain the dependency structures in graph signals.
no code implementations • 6 Jun 2020 • Siu Lun Chau, Javier González, Dino Sejdinovic
We revisit widely used preferential Gaussian processes by Chu et al.(2005) and challenge their modelling assumption that imposes rankability of data items via latent utility function values.
no code implementations • 8 May 2020 • Siu Lun Chau, Mihai Cucuringu, Dino Sejdinovic
We consider spectral approaches to the problem of ranking n players given their incomplete and noisy pairwise comparisons, but revisit this classical problem in light of player covariate information.