no code implementations • 18 Aug 2024 • Yanzhi Chen, Zijing Ou, Adrian Weller, Yingzhen Li
Estimating mutual information (MI) is a fundamental yet challenging task in data science and machine learning.
1 code implementation • 25 Jun 2024 • Carles Balsells-Rodas, Yixin Wang, Pedro A. M. Mediano, Yingzhen Li
Causal discovery in time series is a rapidly evolving field with a wide variety of applications in other areas such as climate science and neuroscience.
no code implementations • 16 Jun 2024 • Zijing Ou, Mingtian Zhang, Andi Zhang, Tim Z. Xiao, Yingzhen Li, David Barber
The probabilistic diffusion model has become highly effective across various domains.
1 code implementation • 21 Mar 2024 • Hee Suk Yoon, Eunseop Yoon, Joshua Tian Jin Tee, Mark Hasegawa-Johnson, Yingzhen Li, Chang D. Yoo
Through a series of observations, we find that the prompt choice significantly affects the calibration in CLIP, where the prompts leading to higher text feature dispersion result in better-calibrated predictions.
no code implementations • 28 Feb 2024 • Laura Manduchi, Kushagra Pandey, Robert Bamler, Ryan Cotterell, Sina Däubener, Sophie Fellenz, Asja Fischer, Thomas Gärtner, Matthias Kirchler, Marius Kloft, Yingzhen Li, Christoph Lippert, Gerard de Melo, Eric Nalisnick, Björn Ommer, Rajesh Ranganath, Maja Rudolph, Karen Ullrich, Guy Van Den Broeck, Julia E Vogt, Yixin Wang, Florian Wenzel, Frank Wood, Stephan Mandt, Vincent Fortuin
The field of deep generative modeling has grown rapidly and consistently over the years.
no code implementations • 1 Feb 2024 • Theodore Papamarkou, Maria Skoularidou, Konstantina Palla, Laurence Aitchison, Julyan Arbel, David Dunson, Maurizio Filippone, Vincent Fortuin, Philipp Hennig, José Miguel Hernández-Lobato, Aliaksandr Hubin, Alexander Immer, Theofanis Karaletsos, Mohammad Emtiyaz Khan, Agustinus Kristiadi, Yingzhen Li, Stephan Mandt, Christopher Nemeth, Michael A. Osborne, Tim G. J. Rudner, David Rügamer, Yee Whye Teh, Max Welling, Andrew Gordon Wilson, Ruqi Zhang
In the current landscape of deep learning research, there is a predominant emphasis on achieving high predictive accuracy in supervised tasks involving large image and language datasets.
no code implementations • 14 Jul 2023 • Tobias Schröder, Zijing Ou, Yingzhen Li, Andrew B. Duncan
Training energy-based models (EBMs) on discrete spaces is challenging because sampling over such spaces can be difficult.
1 code implementation • NeurIPS 2023 • Tobias Schröder, Zijing Ou, Jen Ning Lim, Yingzhen Li, Sebastian J. Vollmer, Andrew B. Duncan
Energy-based models are a simple yet powerful class of probabilistic models, but their widespread adoption has been limited by the computational burden of training them.
1 code implementation • 25 May 2023 • Carles Balsells-Rodas, Yixin Wang, Yingzhen Li
The identifiability of latent variable models has received increasing attention due to its relevance in interpretability and out-of-distribution generalisation.
1 code implementation • 4 Mar 2023 • Wenlong Chen, Yingzhen Li
Transformer models have achieved profound success in prediction tasks in a wide range of applications in natural language processing, speech recognition and computer vision.
1 code implementation • 4 Mar 2023 • Hee Suk Yoon, Joshua Tian Jin Tee, Eunseop Yoon, Sunjae Yoon, Gwangsu Kim, Yingzhen Li, Chang D. Yoo
Studies have shown that modern neural networks tend to be poorly calibrated due to over-confident predictions.
1 code implementation • 21 Feb 2023 • Yanzhi Chen, Weihao Sun, Yingzhen Li, Adrian Weller
The task of infomin learning aims to learn a representation with high utility while being uninformative about a specified target, with the latter achieved by minimising the mutual information between the representation and the target.
no code implementations • 12 Jul 2022 • Harrison Zhu, Carles Balsells Rodas, Yingzhen Li
Sequential VAEs have been successfully considered for many high-dimensional time series modelling problems, with many variant models relying on discrete-time mechanisms such as recurrent neural networks (RNNs).
no code implementations • 11 Jul 2022 • Ryutaro Tanno, Melanie F. Pradier, Aditya Nori, Yingzhen Li
Prediction failures of machine learning models often arise from deficiencies in training data, such as incorrect labels, outliers, and selection biases.
1 code implementation • 3 Mar 2022 • Zijing Ou, Tingyang Xu, Qinliang Su, Yingzhen Li, Peilin Zhao, Yatao Bian
Learning neural set functions becomes increasingly more important in many applications like product recommendation and compound selection in AI-aided drug discovery.
no code implementations • 12 Oct 2021 • Carles Balsells-Rodas, Ruibo Tu, Hedvig Kjellstrom, Yingzhen Li
Causal discovery, i. e., inferring underlying causal relationships from observational data, has been shown to be highly challenging for AI systems.
1 code implementation • 25 Sep 2021 • Thomas Henn, Yasukazu Sakamoto, Clément Jacquet, Shunsuke Yoshizawa, Masamichi Andou, Stephen Tchen, Ryosuke Saga, Hiroyuki Ishihara, Katsuhiko Shimizu, Yingzhen Li, Ryutaro Tanno
We suggest that the quality of the identified failure types can be validated through measuring the intra- and inter-type generalisation after fine-tuning and introduce metrics to compare different subtyping methods.
no code implementations • ICML Workshop INNF 2021 • Wenbo Gong, Yingzhen Li
Specifically, we theoretically prove that DSM (or DSD) is equivalent to the original score matching (or Stein discrepancy) evaluated in the transformed space defined by the normalizing flow, where the diffusion matrix is the inverse of the flow's Jacobian matrix.
no code implementations • NeurIPS 2021 • Hippolyt Ritter, Martin Kukla, Cheng Zhang, Yingzhen Li
Bayesian neural networks and deep ensembles represent two modern paradigms of uncertainty quantification in deep learning.
no code implementations • 12 Apr 2021 • Angus Lamb, Evgeny Saveliev, Yingzhen Li, Sebastian Tschiatschek, Camilla Longden, Simon Woodhead, José Miguel Hernández-Lobato, Richard E. Turner, Pashmina Cameron, Cheng Zhang
While deep learning has obtained state-of-the-art results in many applications, the adaptation of neural network architectures to incorporate new output features remains a challenge, as neural networks are commonly trained to produce a fixed output dimension.
1 code implementation • 5 Feb 2021 • Wenbo Gong, Kaibo Zhang, Yingzhen Li, José Miguel Hernández-Lobato
First, we provide theoretical results stating that the requirement of using optimal slicing directions in the kernelized version of SSD can be relaxed, validating the resulting discrepancy with finite random slicing directions.
1 code implementation • EACL 2021 • Yi Zhu, Ehsan Shareghi, Yingzhen Li, Roi Reichart, Anna Korhonen
Semi-supervised learning through deep generative models and multi-lingual pretraining techniques have orchestrated tremendous success across different areas of NLP.
no code implementations • NeurIPS Workshop LMCA 2020 • Haiyan Yin, Yingzhen Li, Sinno Jialin Pan, Cheng Zhang, Sebastian Tschiatschek
Solving real-life sequential decision making problems under partial observability involves an exploration-exploitation problem.
no code implementations • 28 Oct 2020 • Philip J. Ball, Yingzhen Li, Angus Lamb, Cheng Zhang
We study a setting where the pruning phase is given a time budget, and identify connections between iterative pruning and multiple sleep cycles in humans.
1 code implementation • ACL (RepL4NLP) 2021 • Victor Prokhorov, Yingzhen Li, Ehsan Shareghi, Nigel Collier
It has been long known that sparsity is an effective inductive bias for learning efficient representation of data in vectors with fixed dimensionality, and it has been explored in many areas of representation learning.
no code implementations • 24 Jul 2020 • Chaochao Lu, Richard E. Turner, Yingzhen Li, Nate Kushman
In this paper we provide a firm theoretical interpretation for infinite spatial generation, by drawing connections to spatial stochastic processes.
no code implementations • 6 Jul 2020 • Ruqi Zhang, Yingzhen Li, Christopher De Sa, Sam Devlin, Cheng Zhang
Variational inference (VI) plays an essential role in approximate Bayesian inference due to its computational efficiency and broad applicability.
1 code implementation • ICLR 2021 • Wenbo Gong, Yingzhen Li, José Miguel Hernández-Lobato
Kernelized Stein discrepancy (KSD), though being extensively used in goodness-of-fit tests and model learning, suffers from the curse-of-dimensionality.
no code implementations • 3 May 2020 • Cheng Zhang, Kun Zhang, Yingzhen Li
We present a causal view on the robustness of neural networks against input manipulations, which applies not only to traditional classification tasks but also to general measurement data.
no code implementations • 28 Feb 2020 • Sebastian Lunz, Yingzhen Li, Andrew Fitzgibbon, Nate Kushman
In this paper we introduce the first scalable training technique for 3D generative models from 2D data which utilizes an off-the-shelf non-differentiable renderer.
1 code implementation • NeurIPS 2019 • Maximilian Igl, Kamil Ciosek, Yingzhen Li, Sebastian Tschiatschek, Cheng Zhang, Sam Devlin, Katja Hofmann
We discuss those differences and propose modifications to existing regularization techniques in order to better adapt them to RL.
no code implementations • pproximateinference AABI Symposium 2019 • Chao Ma, Sebastian Tschiatschek, Yingzhen Li, Richard Turner, Jose Miguel Hernandez-Lobato, Cheng Zhang
In this paper, we focused on improving VAEs for real-valued data that has heterogeneous marginal distributions.
1 code implementation • WS 2019 • Victor Prokhorov, Ehsan Shareghi, Yingzhen Li, Mohammad Taher Pilehvar, Nigel Collier
While the explicit constraint naturally avoids posterior collapse, we use it to further understand the significance of the KL term in controlling the information transmitted through the VAE channel.
no code implementations • pproximateinference AABI Symposium 2019 • Ruqi Zhang, Yingzhen Li, Chris De Sa, Sam Devlin, Cheng Zhang
Variational inference (VI) plays an essential role in approximate Bayesian inference due to its computational efficiency and general applicability.
2 code implementations • NeurIPS 2020 • Andrew Y. K. Foong, David R. Burt, Yingzhen Li, Richard E. Turner
While Bayesian neural networks (BNNs) hold the promise of being flexible, well-calibrated statistical models, inference often requires approximations whose consequences are poorly understood.
no code implementations • 27 Jun 2019 • Andrew Y. K. Foong, Yingzhen Li, José Miguel Hernández-Lobato, Richard E. Turner
We describe a limitation in the expressiveness of the predictive uncertainty estimate given by mean-field variational inference (MFVI), a popular approximate inference method for Bayesian neural networks.
no code implementations • NAACL 2019 • Ehsan Shareghi, Yingzhen Li, Yi Zhu, Roi Reichart, Anna Korhonen
While neural dependency parsers provide state-of-the-art accuracy for several languages, they still rely on large amounts of costly labeled training data.
2 code implementations • 7 May 2019 • Hiske Overweg, Anna-Lena Popkes, Ari Ercole, Yingzhen Li, José Miguel Hernández-Lobato, Yordan Zaykov, Cheng Zhang
However, flexible tools such as artificial neural networks (ANNs) suffer from a lack of interpretability limiting their acceptability to clinicians.
1 code implementation • ICLR 2019 • Wenbo Gong, Yingzhen Li, José Miguel Hernández-Lobato
Stochastic gradient Markov chain Monte Carlo (SG-MCMC) has become increasingly popular for simulating posterior samples in large-scale Bayesian modeling.
1 code implementation • 6 Jun 2018 • Chao Ma, Yingzhen Li, José Miguel Hernández-Lobato
We introduce the implicit processes (IPs), a stochastic process that places implicitly defined multivariate distributions over any finite collections of random variables.
3 code implementations • ICML 2018 • Yingzhen Li, Stephan Mandt
This architecture gives us partial control over generating content and dynamics by conditioning on either one of these sets of features.
1 code implementation • 19 Feb 2018 • Yingzhen Li, John Bradshaw, Yash Sharma
There is a rising interest in studying the robustness of deep neural network classifiers against adversaries, with both advanced attack and defence techniques being actively developed.
8 code implementations • ICLR 2018 • Cuong V. Nguyen, Yingzhen Li, Thang D. Bui, Richard E. Turner
This paper develops variational continual learning (VCL), a simple but general framework for continual learning that fuses online variational inference (VI) and recent advances in Monte Carlo VI for neural networks.
1 code implementation • ICLR 2018 • Yingzhen Li, Richard E. Turner
Implicit models, which allow for the generation of samples but not for point-wise evaluation of probabilities, are omnipresent in real-world problems tackled by machine learning and a hot topic of current research.
1 code implementation • ICML 2017 • Yingzhen Li, Yarin Gal
To obtain uncertainty estimates with real-world Bayesian deep learning models, practical inference approximations are needed.
no code implementations • 27 Feb 2017 • Yingzhen Li, Richard E. Turner, Qiang Liu
We propose a novel approximate inference algorithm that approximates a target distribution by amortising the dynamics of a user-selected MCMC sampler.
no code implementations • 12 Feb 2016 • Thang D. Bui, Daniel Hernández-Lobato, Yingzhen Li, José Miguel Hernández-Lobato, Richard E. Turner
Deep Gaussian processes (DGPs) are multi-layer hierarchical generalisations of Gaussian processes (GPs) and are formally equivalent to neural networks with multiple, infinitely wide hidden layers.
2 code implementations • NeurIPS 2016 • Yingzhen Li, Richard E. Turner
This paper introduces the variational R\'enyi bound (VR) that extends traditional variational inference to R\'enyi's alpha-divergences.
no code implementations • 11 Nov 2015 • Thang D. Bui, José Miguel Hernández-Lobato, Yingzhen Li, Daniel Hernández-Lobato, Richard E. Turner
Deep Gaussian processes (DGPs) are multi-layer hierarchical generalisations of Gaussian processes (GPs) and are formally equivalent to neural networks with multiple, infinitely wide hidden layers.
no code implementations • 10 Nov 2015 • Daniel Hernández-Lobato, José Miguel Hernández-Lobato, Yingzhen Li, Thang Bui, Richard E. Turner
A method for large scale Gaussian process classification has been recently proposed based on expectation propagation (EP).
3 code implementations • 10 Nov 2015 • José Miguel Hernández-Lobato, Yingzhen Li, Mark Rowland, Daniel Hernández-Lobato, Thang Bui, Richard E. Turner
Black-box alpha (BB-$\alpha$) is a new approximate inference method based on the minimization of $\alpha$-divergences.
no code implementations • NeurIPS 2015 • Yingzhen Li, Jose Miguel Hernandez-Lobato, Richard E. Turner
Expectation propagation (EP) is a deterministic approximation algorithm that is often used to perform approximate Bayesian parameter learning.
no code implementations • 21 Aug 2012 • Yingzhen Li, Ye Zhang
Precise recommendation of followers helps in improving the user experience and maintaining the prosperity of twitter and microblog platforms.