Search Results for author: Maximilian Ilse

Found 10 papers, 4 papers with code

RAD-DINO: Exploring Scalable Medical Image Encoders Beyond Text Supervision

no code implementations19 Jan 2024 Fernando Pérez-García, Harshita Sharma, Sam Bond-Taylor, Kenza Bouzid, Valentina Salvatelli, Maximilian Ilse, Shruthi Bannur, Daniel C. Castro, Anton Schwaighofer, Matthew P. Lungren, Maria Wetscherek, Noel Codella, Stephanie L. Hyland, Javier Alvarez-Valle, Ozan Oktay

We introduce RAD-DINO, a biomedical image encoder pre-trained solely on unimodal biomedical imaging data that obtains similar or greater performance than state-of-the-art biomedical language supervised models on a diverse range of benchmarks.

Semantic Segmentation

Combining Interventional and Observational Data Using Causal Reductions

1 code implementation8 Mar 2021 Maximilian Ilse, Patrick Forré, Max Welling, Joris M. Mooij

Second, for continuous variables and assuming a linear-Gaussian model, we derive equality constraints for the parameters of the observational and interventional distributions.

Causal Inference

Problems using deep generative models for probabilistic audio source separation

no code implementations NeurIPS Workshop ICBINB 2020 Maurice Frank, Maximilian Ilse

Recent advancements in deep generative modeling make it possible to learn prior distributions from complex data that subsequently can be used for Bayesian inference.

Audio Source Separation Bayesian Inference

Selecting Data Augmentation for Simulating Interventions

1 code implementation4 May 2020 Maximilian Ilse, Jakub M. Tomczak, Patrick Forré

We argue that causal concepts can be used to explain the success of data augmentation by describing how they can weaken the spurious correlation between the observed domains and the task labels.

Data Augmentation Domain Generalization

DIVA: Domain Invariant Variational Autoencoders

3 code implementations24 May 2019 Maximilian Ilse, Jakub M. Tomczak, Christos Louizos, Max Welling

We consider the problem of domain generalization, namely, how to learn representations given data from a set of domains that generalize to data from a previously unseen domain.

Domain Generalization Rotated MNIST

DIVA: Domain Invariant Variational Autoencoder

no code implementations ICLR Workshop DeepGenStruct 2019 Maximilian Ilse, Jakub M. Tomczak, Christos Louizos, Max Welling

We consider the problem of domain generalization, namely, how to learn representations given data from a set of domains that generalize to data from a previously unseen domain.

Domain Generalization Rotated MNIST

Cannot find the paper you are looking for? You can Submit a new open access paper.