Search Results for author: Yingzhen Li

Found 50 papers, 23 papers with code

C-TPT: Calibrated Test-Time Prompt Tuning for Vision-Language Models via Text Feature Dispersion

no code implementations21 Mar 2024 Hee Suk Yoon, Eunseop Yoon, Joshua Tian Jin Tee, Mark Hasegawa-Johnson, Yingzhen Li, Chang D. Yoo

Through a series of observations, we find that the prompt choice significantly affects the calibration in CLIP, where the prompts leading to higher text feature dispersion result in better-calibrated predictions.

Test-time Adaptation

Training Discrete Energy-Based Models with Energy Discrepancy

no code implementations14 Jul 2023 Tobias Schröder, Zijing Ou, Yingzhen Li, Andrew B. Duncan

Training energy-based models (EBMs) on discrete spaces is challenging because sampling over such spaces can be difficult.

Energy Discrepancies: A Score-Independent Loss for Energy-Based Models

1 code implementation NeurIPS 2023 Tobias Schröder, Zijing Ou, Jen Ning Lim, Yingzhen Li, Sebastian J. Vollmer, Andrew B. Duncan

Energy-based models are a simple yet powerful class of probabilistic models, but their widespread adoption has been limited by the computational burden of training them.

On the Identifiability of Switching Dynamical Systems

no code implementations25 May 2023 Carles Balsells-Rodas, Yixin Wang, Yingzhen Li

In the realm of interpretability and out-of-distribution generalisation, the identifiability of latent variable models has emerged as a captivating field of inquiry.

Causal Discovery Time Series

ESD: Expected Squared Difference as a Tuning-Free Trainable Calibration Measure

1 code implementation4 Mar 2023 Hee Suk Yoon, Joshua Tian Jin Tee, Eunseop Yoon, Sunjae Yoon, Gwangsu Kim, Yingzhen Li, Chang D. Yoo

Studies have shown that modern neural networks tend to be poorly calibrated due to over-confident predictions.

Calibrating Transformers via Sparse Gaussian Processes

1 code implementation4 Mar 2023 Wenlong Chen, Yingzhen Li

Transformer models have achieved profound success in prediction tasks in a wide range of applications in natural language processing, speech recognition and computer vision.

Bayesian Inference Gaussian Processes +3

Scalable Infomin Learning

1 code implementation21 Feb 2023 Yanzhi Chen, Weihao Sun, Yingzhen Li, Adrian Weller

The task of infomin learning aims to learn a representation with high utility while being uninformative about a specified target, with the latter achieved by minimising the mutual information between the representation and the target.

Domain Adaptation Fairness +1

Markovian Gaussian Process Variational Autoencoders

no code implementations12 Jul 2022 Harrison Zhu, Carles Balsells Rodas, Yingzhen Li

Sequential VAEs have been successfully considered for many high-dimensional time series modelling problems, with many variant models relying on discrete-time mechanisms such as recurrent neural networks (RNNs).

Time Series Time Series Analysis

Repairing Neural Networks by Leaving the Right Past Behind

no code implementations11 Jul 2022 Ryutaro Tanno, Melanie F. Pradier, Aditya Nori, Yingzhen Li

Prediction failures of machine learning models often arise from deficiencies in training data, such as incorrect labels, outliers, and selection biases.

Continual Learning

Learning Neural Set Functions Under the Optimal Subset Oracle

1 code implementation3 Mar 2022 Zijing Ou, Tingyang Xu, Qinliang Su, Yingzhen Li, Peilin Zhao, Yatao Bian

Learning neural set functions becomes increasingly more important in many applications like product recommendation and compound selection in AI-aided drug discovery.

Anomaly Detection Drug Discovery +2

Causal Discovery from Conditionally Stationary Time Series

no code implementations12 Oct 2021 Carles Balsells-Rodas, Ruibo Tu, Hedvig Kjellstrom, Yingzhen Li

Causal discovery, i. e., inferring underlying causal relationships from observational data, has been shown to be highly challenging for AI systems.

Causal Discovery Causal Inference +3

A Principled Approach to Failure Analysis and Model Repairment: Demonstration in Medical Imaging

1 code implementation25 Sep 2021 Thomas Henn, Yasukazu Sakamoto, Clément Jacquet, Shunsuke Yoshizawa, Masamichi Andou, Stephen Tchen, Ryosuke Saga, Hiroyuki Ishihara, Katsuhiko Shimizu, Yingzhen Li, Ryutaro Tanno

We suggest that the quality of the identified failure types can be validated through measuring the intra- and inter-type generalisation after fine-tuning and introduce metrics to compare different subtyping methods.

object-detection Object Detection

Interpreting diffusion score matching using normalizing flow

no code implementations ICML Workshop INNF 2021 Wenbo Gong, Yingzhen Li

Specifically, we theoretically prove that DSM (or DSD) is equivalent to the original score matching (or Stein discrepancy) evaluated in the transformed space defined by the normalizing flow, where the diffusion matrix is the inverse of the flow's Jacobian matrix.

Sparse Uncertainty Representation in Deep Learning with Inducing Weights

no code implementations NeurIPS 2021 Hippolyt Ritter, Martin Kukla, Cheng Zhang, Yingzhen Li

Bayesian neural networks and deep ensembles represent two modern paradigms of uncertainty quantification in deep learning.

Uncertainty Quantification

Contextual HyperNetworks for Novel Feature Adaptation

no code implementations12 Apr 2021 Angus Lamb, Evgeny Saveliev, Yingzhen Li, Sebastian Tschiatschek, Camilla Longden, Simon Woodhead, José Miguel Hernández-Lobato, Richard E. Turner, Pashmina Cameron, Cheng Zhang

While deep learning has obtained state-of-the-art results in many applications, the adaptation of neural network architectures to incorporate new output features remains a challenge, as neural networks are commonly trained to produce a fixed output dimension.

Few-Shot Learning Imputation +1

Active Slices for Sliced Stein Discrepancy

1 code implementation5 Feb 2021 Wenbo Gong, Kaibo Zhang, Yingzhen Li, José Miguel Hernández-Lobato

First, we provide theoretical results stating that the requirement of using optimal slicing directions in the kernelized version of SSD can be relaxed, validating the resulting discrepancy with finite random slicing directions.

Combining Deep Generative Models and Multi-lingual Pretraining for Semi-supervised Document Classification

1 code implementation EACL 2021 Yi Zhu, Ehsan Shareghi, Yingzhen Li, Roi Reichart, Anna Korhonen

Semi-supervised learning through deep generative models and multi-lingual pretraining techniques have orchestrated tremendous success across different areas of NLP.

Classification Document Classification +1

A Study on Efficiency in Continual Learning Inspired by Human Learning

no code implementations28 Oct 2020 Philip J. Ball, Yingzhen Li, Angus Lamb, Cheng Zhang

We study a setting where the pruning phase is given a time budget, and identify connections between iterative pruning and multiple sleep cycles in humans.

Continual Learning

Learning Sparse Sentence Encoding without Supervision: An Exploration of Sparsity in Variational Autoencoders

1 code implementation ACL (RepL4NLP) 2021 Victor Prokhorov, Yingzhen Li, Ehsan Shareghi, Nigel Collier

It has been long known that sparsity is an effective inductive bias for learning efficient representation of data in vectors with fixed dimensionality, and it has been explored in many areas of representation learning.

Inductive Bias Representation Learning +3

Interpreting Spatially Infinite Generative Models

no code implementations24 Jul 2020 Chaochao Lu, Richard E. Turner, Yingzhen Li, Nate Kushman

In this paper we provide a firm theoretical interpretation for infinite spatial generation, by drawing connections to spatial stochastic processes.

Generative Adversarial Network Texture Synthesis

Meta-Learning Divergences of Variational Inference

no code implementations6 Jul 2020 Ruqi Zhang, Yingzhen Li, Christopher De Sa, Sam Devlin, Cheng Zhang

Variational inference (VI) plays an essential role in approximate Bayesian inference due to its computational efficiency and broad applicability.

Bayesian Inference Computational Efficiency +4

Sliced Kernelized Stein Discrepancy

1 code implementation ICLR 2021 Wenbo Gong, Yingzhen Li, José Miguel Hernández-Lobato

Kernelized Stein discrepancy (KSD), though being extensively used in goodness-of-fit tests and model learning, suffers from the curse-of-dimensionality.

A Causal View on Robustness of Neural Networks

no code implementations3 May 2020 Cheng Zhang, Kun Zhang, Yingzhen Li

We present a causal view on the robustness of neural networks against input manipulations, which applies not only to traditional classification tasks but also to general measurement data.

Data Augmentation

Inverse Graphics GAN: Learning to Generate 3D Shapes from Unstructured 2D Data

no code implementations28 Feb 2020 Sebastian Lunz, Yingzhen Li, Andrew Fitzgibbon, Nate Kushman

In this paper we introduce the first scalable training technique for 3D generative models from 2D data which utilizes an off-the-shelf non-differentiable renderer.

On the Importance of the Kullback-Leibler Divergence Term in Variational Autoencoders for Text Generation

1 code implementation WS 2019 Victor Prokhorov, Ehsan Shareghi, Yingzhen Li, Mohammad Taher Pilehvar, Nigel Collier

While the explicit constraint naturally avoids posterior collapse, we use it to further understand the significance of the KL term in controlling the information transmitted through the VAE channel.

Text Generation

Meta-Learning for Variational Inference

no code implementations pproximateinference AABI Symposium 2019 Ruqi Zhang, Yingzhen Li, Chris De Sa, Sam Devlin, Cheng Zhang

Variational inference (VI) plays an essential role in approximate Bayesian inference due to its computational efficiency and general applicability.

Bayesian Inference Computational Efficiency +4

On the Expressiveness of Approximate Inference in Bayesian Neural Networks

2 code implementations NeurIPS 2020 Andrew Y. K. Foong, David R. Burt, Yingzhen Li, Richard E. Turner

While Bayesian neural networks (BNNs) hold the promise of being flexible, well-calibrated statistical models, inference often requires approximations whose consequences are poorly understood.

Active Learning Bayesian Inference +3

'In-Between' Uncertainty in Bayesian Neural Networks

no code implementations27 Jun 2019 Andrew Y. K. Foong, Yingzhen Li, José Miguel Hernández-Lobato, Richard E. Turner

We describe a limitation in the expressiveness of the predictive uncertainty estimate given by mean-field variational inference (MFVI), a popular approximate inference method for Bayesian neural networks.

Active Learning Bayesian Optimisation +1

Bayesian Learning for Neural Dependency Parsing

no code implementations NAACL 2019 Ehsan Shareghi, Yingzhen Li, Yi Zhu, Roi Reichart, Anna Korhonen

While neural dependency parsers provide state-of-the-art accuracy for several languages, they still rely on large amounts of costly labeled training data.

Dependency Parsing POS +2

Interpretable Outcome Prediction with Sparse Bayesian Neural Networks in Intensive Care

2 code implementations7 May 2019 Hiske Overweg, Anna-Lena Popkes, Ari Ercole, Yingzhen Li, José Miguel Hernández-Lobato, Yordan Zaykov, Cheng Zhang

However, flexible tools such as artificial neural networks (ANNs) suffer from a lack of interpretability limiting their acceptability to clinicians.

Decision Making feature selection +1

Meta-Learning for Stochastic Gradient MCMC

1 code implementation ICLR 2019 Wenbo Gong, Yingzhen Li, José Miguel Hernández-Lobato

Stochastic gradient Markov chain Monte Carlo (SG-MCMC) has become increasingly popular for simulating posterior samples in large-scale Bayesian modeling.

Efficient Exploration Meta-Learning +1

Variational Implicit Processes

1 code implementation6 Jun 2018 Chao Ma, Yingzhen Li, José Miguel Hernández-Lobato

We introduce the implicit processes (IPs), a stochastic process that places implicitly defined multivariate distributions over any finite collections of random variables.

Gaussian Processes Stochastic Optimization

Disentangled Sequential Autoencoder

3 code implementations ICML 2018 Yingzhen Li, Stephan Mandt

This architecture gives us partial control over generating content and dynamics by conditioning on either one of these sets of features.

Video Compression

Are Generative Classifiers More Robust to Adversarial Attacks?

1 code implementation19 Feb 2018 Yingzhen Li, John Bradshaw, Yash Sharma

There is a rising interest in studying the robustness of deep neural network classifiers against adversaries, with both advanced attack and defence techniques being actively developed.

Adversarial Defense Adversarial Robustness

Variational Continual Learning

8 code implementations ICLR 2018 Cuong V. Nguyen, Yingzhen Li, Thang D. Bui, Richard E. Turner

This paper develops variational continual learning (VCL), a simple but general framework for continual learning that fuses online variational inference (VI) and recent advances in Monte Carlo VI for neural networks.

Continual Learning Variational Inference

Gradient Estimators for Implicit Models

1 code implementation ICLR 2018 Yingzhen Li, Richard E. Turner

Implicit models, which allow for the generation of samples but not for point-wise evaluation of probabilities, are omnipresent in real-world problems tackled by machine learning and a hot topic of current research.

Image Generation Meta-Learning

Dropout Inference in Bayesian Neural Networks with Alpha-divergences

1 code implementation ICML 2017 Yingzhen Li, Yarin Gal

To obtain uncertainty estimates with real-world Bayesian deep learning models, practical inference approximations are needed.

Variational Inference

Approximate Inference with Amortised MCMC

no code implementations27 Feb 2017 Yingzhen Li, Richard E. Turner, Qiang Liu

We propose a novel approximate inference algorithm that approximates a target distribution by amortising the dynamics of a user-selected MCMC sampler.

Deep Gaussian Processes for Regression using Approximate Expectation Propagation

no code implementations12 Feb 2016 Thang D. Bui, Daniel Hernández-Lobato, Yingzhen Li, José Miguel Hernández-Lobato, Richard E. Turner

Deep Gaussian processes (DGPs) are multi-layer hierarchical generalisations of Gaussian processes (GPs) and are formally equivalent to neural networks with multiple, infinitely wide hidden layers.

Gaussian Processes regression

Rényi Divergence Variational Inference

2 code implementations NeurIPS 2016 Yingzhen Li, Richard E. Turner

This paper introduces the variational R\'enyi bound (VR) that extends traditional variational inference to R\'enyi's alpha-divergences.

Variational Inference

Training Deep Gaussian Processes using Stochastic Expectation Propagation and Probabilistic Backpropagation

no code implementations11 Nov 2015 Thang D. Bui, José Miguel Hernández-Lobato, Yingzhen Li, Daniel Hernández-Lobato, Richard E. Turner

Deep Gaussian processes (DGPs) are multi-layer hierarchical generalisations of Gaussian processes (GPs) and are formally equivalent to neural networks with multiple, infinitely wide hidden layers.

Gaussian Processes

Black-box $α$-divergence Minimization

3 code implementations10 Nov 2015 José Miguel Hernández-Lobato, Yingzhen Li, Mark Rowland, Daniel Hernández-Lobato, Thang Bui, Richard E. Turner

Black-box alpha (BB-$\alpha$) is a new approximate inference method based on the minimization of $\alpha$-divergences.

General Classification regression

Stochastic Expectation Propagation

no code implementations NeurIPS 2015 Yingzhen Li, Jose Miguel Hernandez-Lobato, Richard E. Turner

Expectation propagation (EP) is a deterministic approximation algorithm that is often used to perform approximate Bayesian parameter learning.

Variational Inference

Generating ordered list of Recommended Items: a Hybrid Recommender System of Microblog

no code implementations21 Aug 2012 Yingzhen Li, Ye Zhang

Precise recommendation of followers helps in improving the user experience and maintaining the prosperity of twitter and microblog platforms.

Recommendation Systems

Cannot find the paper you are looking for? You can Submit a new open access paper.