no code implementations • Findings (EMNLP) 2021 • Ping Yu, Yang Zhao, Chunyuan Li, Changyou Chen
To overcome this issue, we propose a graph-based method to extract attribute content and attribute-independent content from input sentences in the YELP dataset and IMDB dataset.
1 code implementation • 20 Dec 2024 • Shijie Zhou, Ruiyi Zhang, Yufan Zhou, Changyou Chen
Large multimodal models still struggle with text-rich images because of inadequate training data.
no code implementations • 15 Nov 2024 • Shijie Zhou, Huaisheng Zhu, Rohan Sharma, Ruiyi Zhang, Kaiyi Ji, Changyou Chen
Diffusion models have emerged as a powerful foundation model for visual generation.
no code implementations • 2 Nov 2024 • Jian Chen, Ruiyi Zhang, Yufan Zhou, Tong Yu, Franck Dernoncourt, Jiuxiang Gu, Ryan A. Rossi, Changyou Chen, Tong Sun
In this work, we present a novel framework named LoRA-Contextualizing Adaptation of Large multimodal models (LoCAL), which broadens the capabilities of any LMM to support long-document understanding.
1 code implementation • 9 Oct 2024 • Jian Chen, Ruiyi Zhang, Yufan Zhou, Jennifer Healey, Jiuxiang Gu, Zhiqiang Xu, Changyou Chen
Automatic generation of graphical layouts is crucial for many real-world applications, including designing posters, flyers, advertisements, and graphical user interfaces.
1 code implementation • 26 Aug 2024 • Jian Chen, Ruiyi Zhang, Yufan Zhou, Ryan Rossi, Jiuxiang Gu, Changyou Chen
Large multimodal models (LMMs) have demonstrated impressive capabilities in understanding various types of image, including text-rich images.
no code implementations • 27 Jul 2024 • Ruiyi Zhang, Yufan Zhou, Jian Chen, Jiuxiang Gu, Changyou Chen, Tong Sun
Large multimodal language models have demonstrated impressive capabilities in understanding and manipulating images.
no code implementations • 24 Jul 2024 • Changyou Chen, Han Ding, Bunyamin Sisman, Yi Xu, Ouye Xie, Benjamin Z. Yao, Son Dinh Tran, Belinda Zeng
In this paper, we propose a principled way to define a diffusion model by constructing a unified multi-modal diffusion model in a common diffusion space.
1 code implementation • 22 Jul 2024 • Jingchen Sun, Rohan Sharma, Vishnu Suresh Lokhande, Changyou Chen
The experiment on four different prompt tuning structures consistently shows the improvement of our method, with increases of up to $6. 1\%$ in the Base-to-Novel generalization task, $5. 8\%$ in the group robustness task, and $2. 7\%$ in the out-of-distribution tasks.
1 code implementation • CVPR 2024 • Ruiyi Zhang, Yanzhe Zhang, Jian Chen, Yufan Zhou, Jiuxiang Gu, Changyou Chen, Tong Sun
In this work, we introduce TRINS: a Text-Rich image INStruction dataset, with the objective of enhancing the reading ability of the multimodal large language model.
no code implementations • 2 May 2024 • Somesh Singh, Harini S I, Yaman K Singla, Veeky Baths, Rajiv Ratn Shah, Changyou Chen, Balaji Krishnamurthy
Specifically, we show that training LLMs to predict the receiver behavior of likes and comments improves the LLM's performance on a wide variety of downstream content understanding tasks.
no code implementations • 10 Feb 2024 • Rohan Sharma, Shijie Zhou, Kaiyi Ji, Changyou Chen
We consider the scenario of two networks, the attacker $\mathbf{A}$ and the trained defender $\mathbf{D}$ pitted against each other in an adversarial objective, wherein the attacker aims at teasing out the information of the data to be unlearned in order to infer membership, and the defender unlearns to defend the network against the attack, whilst preserving its general performance.
1 code implementation • 7 Feb 2024 • Jian Chen, Ruiyi Zhang, Yufan Zhou, Rajiv Jain, Zhiqiang Xu, Ryan Rossi, Changyou Chen
Controllable layout generation refers to the process of creating a plausible visual arrangement of elements within a graphic design (e. g., document and web designs) with constraints representing design intentions.
no code implementations • 18 Nov 2023 • Varun Khurana, Yaman K Singla, Jayakumar Subramanian, Rajiv Ratn Shah, Changyou Chen, Zhiqiang Xu, Balaji Krishnamurthy
We show that BoigLLM outperforms 13x larger models such as GPT-3. 5 and GPT-4 in this task, demonstrating that while these state-of-the-art models can understand images, they lack information on how these images perform in the real world.
no code implementations • 2 Nov 2023 • Ngoc Dang Nguyen, Wei Tan, Lan Du, Wray Buntine, Richard Beare, Changyou Chen
Named entity recognition (NER), a task that identifies and categorizes named entities such as persons or organizations from text, is traditionally framed as a multi-class classification problem.
no code implementations • 1 Sep 2023 • Harini SI, Somesh Singh, Yaman K Singla, Aanisha Bhattacharyya, Veeky Baths, Changyou Chen, Rajiv Ratn Shah, Balaji Krishnamurthy
Finally, with the intent of memorable ad generation, we present a scalable method to build a high-quality memorable ad generation model by leveraging automatically annotated data.
no code implementations • 1 Sep 2023 • Ashmit Khandelwal, Aditya Agrawal, Aanisha Bhattacharyya, Yaman K Singla, Somesh Singh, Uttaran Bhattacharya, Ishita Dasgupta, Stefano Petrangeli, Rajiv Ratn Shah, Changyou Chen, Balaji Krishnamurthy
We call these models Large Content and Behavior Models (LCBMs).
no code implementations • 13 Jun 2023 • Wei Jiang, Jiayu Qin, Lingyu Wu, Changyou Chen, Tianbao Yang, Lijun Zhang
Learning unnormalized statistical models (e. g., energy-based models) is computationally challenging due to the complexity of handling the partition function.
1 code implementation • NeurIPS 2023 • Jian Chen, Ruiyi Zhang, Tong Yu, Rohan Sharma, Zhiqiang Xu, Tong Sun, Changyou Chen
Remarkably, by incorporating conditional information from the powerful CLIP model, our method can boost the current SOTA accuracy by 10-20 absolute points in many cases.
Ranked #1 on Image Classification on Food-101N (using extra training data)
1 code implementation • 16 May 2023 • Aanisha Bhattacharya, Yaman K Singla, Balaji Krishnamurthy, Rajiv Ratn Shah, Changyou Chen
Multimedia content, such as advertisements and story videos, exhibit a rich blend of creativity and multiple modalities.
1 code implementation • 24 Mar 2023 • Jingchen Sun, Jiayu Qin, Zihao Lin, Changyou Chen
To address this issue, efficient adaptation methods such as prompt tuning have been proposed.
no code implementations • CVPR 2023 • Qian Jiang, Changyou Chen, Han Zhao, Liqun Chen, Qing Ping, Son Dinh Tran, Yi Xu, Belinda Zeng, Trishul Chilimbi
Hence we advocate that the key of better performance lies in meaningful latent modality structures instead of perfect modality alignment.
1 code implementation • 21 Dec 2022 • Zihao He, Weituo Hao, Wei-Tsung Lu, Changyou Chen, Kristina Lerman, Xuchen Song
Music captioning has gained significant attention in the wake of the rising prominence of streaming media platforms.
no code implementations • 9 Dec 2022 • Ngoc Dang Nguyen, Wei Tan, Wray Buntine, Richard Beare, Changyou Chen, Lan Du
To the best of our knowledge, this is the first work that brings AUC maximization to the NER setting.
Low Resource Named Entity Recognition named-entity-recognition +2
1 code implementation • CVPR 2023 • Yufan Zhou, Bingchen Liu, Yizhe Zhu, Xiao Yang, Changyou Chen, Jinhui Xu
Unlike the baseline diffusion model used in DALL-E 2, our method seamlessly encodes prior knowledge of the pre-trained CLIP model in its diffusion process by designing a new initialization distribution and a new transition step of the diffusion.
Ranked #3 on Text-to-Image Generation on Multi-Modal-CelebA-HQ
no code implementations • 11 Nov 2022 • Ngoc Dang Nguyen, Lan Du, Wray Buntine, Changyou Chen, Richard Beare
Domain adaptation is an effective solution to data scarcity in low-resource scenarios.
no code implementations • 25 Oct 2022 • Yufan Zhou, Chunyuan Li, Changyou Chen, Jianfeng Gao, Jinhui Xu
The low requirement of the proposed method yields high flexibility and usability: it can be beneficial to a wide range of settings, including the few-shot, semi-supervised and fully-supervised learning; it can be applied on different models including generative adversarial networks (GANs) and diffusion models.
no code implementations • 30 Sep 2022 • Jianyi Zhang, Ang Li, Minxue Tang, Jingwei Sun, Xiang Chen, Fan Zhang, Changyou Chen, Yiran Chen, Hai Li
Based on this measure, we also design a computation-efficient client sampling strategy, such that the actively selected clients will generate a more class-balanced grouped dataset with theoretical guarantees.
1 code implementation • 20 Aug 2022 • Yaman Kumar Singla, Rajat Jha, Arunim Gupta, Milan Aggarwal, Aditya Garg, Tushar Malyan, Ayush Bhardwaj, Rajiv Ratn Shah, Balaji Krishnamurthy, Changyou Chen
Motivated by persuasion literature in social psychology and marketing, we introduce an extensive vocabulary of persuasion strategies and build the first ad image corpus annotated with persuasion strategies.
no code implementations • 18 Jul 2022 • Ping Yu, Wei Wang, Chunyuan Li, Ruiyi Zhang, Zhanpeng Jin, Changyou Chen
Significantly, it can even outperform the time- and resource-consuming fine-tuning method on sentiment classification tasks.
no code implementations • CVPR 2022 • Yufan Zhou, Ruiyi Zhang, Changyou Chen, Chunyuan Li, Chris Tensmeyer, Tong Yu, Jiuxiang Gu, Jinhui Xu, Tong Sun
One of the major challenges in training text-to-image generation models is the need of a large number of high-quality text-image pairs.
no code implementations • 7 Dec 2021 • Yufan Zhou, Chunyuan Li, Changyou Chen, Jinhui Xu
With the rapidly growing model complexity and data volume, training deep generative models (DGMs) for better performance has becoming an increasingly more important challenge.
3 code implementations • 27 Nov 2021 • Yufan Zhou, Ruiyi Zhang, Changyou Chen, Chunyuan Li, Chris Tensmeyer, Tong Yu, Jiuxiang Gu, Jinhui Xu, Tong Sun
One of the major challenges in training text-to-image generation models is the need of a large number of high-quality image-text pairs.
Ranked #2 on Text-to-Image Generation on Multi-Modal-CelebA-HQ
1 code implementation • 17 Nov 2021 • Yaman Kumar Singla, Sriram Krishna, Rajiv Ratn Shah, Changyou Chen
Automated Scoring (AS), the natural language processing task of scoring essays and speeches in an educational testing setting, is growing in popularity and being deployed across contexts from government examinations to companies providing language proficiency services.
no code implementations • 13 Oct 2021 • Anuj Saraswat, Mehar Bhatia, Yaman Kumar Singla, Changyou Chen, Rajiv Ratn Shah
Recent studies in speech perception have been closely linked to fields of cognitive psychology, phonology, and phonetics in linguistics.
no code implementations • CVPR 2022 • Yang Zhao, Yu-Chuan Su, Chun-Te Chu, Yandong Li, Marius Renn, Yukun Zhu, Changyou Chen, Xuhui Jia
While existing approaches for face restoration make significant progress in generating high-quality faces, they often fail to preserve facial features and cannot authentically reconstruct the faces.
1 code implementation • 25 Sep 2021 • Swapnil Parekh, Yaman Singla Kumar, Somesh Singh, Changyou Chen, Balaji Krishnamurthy, Rajiv Ratn Shah
It is well known that natural language models are vulnerable to adversarial attacks, which are mostly input-specific in nature.
no code implementations • 24 Sep 2021 • Yaman Kumar Singla, Swapnil Parekh, Somesh Singh, Junyi Jessy Li, Rajiv Ratn Shah, Changyou Chen
This is in stark contrast to recent probing studies on pre-trained representation learning models, which show that rich linguistic features such as parts-of-speech and morphology are encoded by them.
no code implementations • 30 Aug 2021 • Yaman Kumar Singla, Avykat Gupta, Shaurya Bagga, Changyou Chen, Balaji Krishnamurthy, Rajiv Ratn Shah
In our technique, we take advantage of the fact that oral proficiency tests rate multiple responses for a candidate.
6 code implementations • ACL 2021 • Zijing Ou, Qinliang Su, Jianxing Yu, Bang Liu, Jingwen Wang, Ruihui Zhao, Changyou Chen, Yefeng Zheng
With the need of fast retrieval speed and small memory footprint, document hashing has been playing a crucial role in large-scale information retrieval.
1 code implementation • 13 May 2021 • Zexuan Qiu, Qinliang Su, Zijing Ou, Jianxing Yu, Changyou Chen
Many unsupervised hashing methods are implicitly established on the idea of reconstructing the input data, which basically encourages the hashing codes to retain as much information of original data as possible.
no code implementations • 10 May 2021 • Yufan Zhou, Changyou Chen, Jinhui Xu
Learning high-dimensional distributions is an important yet challenging problem in machine learning with applications in various domains.
no code implementations • 27 Apr 2021 • Weituo Hao, Mostafa El-Khamy, Jungwon Lee, Jianyi Zhang, Kevin J Liang, Changyou Chen, Lawrence Carin
Federated learning has emerged as an important distributed learning paradigm, where a server aggregates a global model from many client-trained models while having no access to the client data.
no code implementations • 7 Feb 2021 • Yufan Zhou, Zhenyi Wang, Jiayi Xian, Changyou Chen, Jinhui Xu
We achieve this goal by 1) replacing the adaptation with a fast-adaptive regularizer in the RKHS; and 2) solving the adaptation analytically based on the NTK theory.
1 code implementation • 4 Jan 2021 • Le Fang, Tao Zeng, Chaochun Liu, Liefeng Bo, Wen Dong, Changyou Chen
Our paper is among the first ones by our knowledge to propose a model and to create datasets for the task of "outline to story".
2 code implementations • 4 Jan 2021 • Le Fang, Tao Zeng, Chaochun Liu, Liefeng Bo, Wen Dong, Changyou Chen
In this paper, we advocate to revive latent variable modeling, essentially the power of representation learning, in the era of Transformers to enhance controllability without hurting state-of-the-art generation effectiveness.
no code implementations • 2 Jan 2021 • Ping Yu, Ruiyi Zhang, Yang Zhao, Yizhe Zhang, Chunyuan Li, Changyou Chen
Data augmentation has been widely used to improve deep neural networks in many research fields, such as computer vision.
no code implementations • 2 Jan 2021 • Jui Shah, Yaman Kumar Singla, Changyou Chen, Rajiv Ratn Shah
In recent times, BERT based transformer models have become an inseparable part of the 'tech stack' of text processing models.
no code implementations • ICLR 2021 • Yufan Zhou, Zhenyi Wang, Jiayi Xian, Changyou Chen, Jinhui Xu
Within this paradigm, we introduce two meta learning algorithms in RKHS, which no longer need an explicit inner-loop adaptation as in the MAML framework.
no code implementations • 1 Jan 2021 • Zhenyi Wang, Tiehang Duan, Donglin Zhan, Changyou Chen
However, a natural generalization to the sequential domain setting to avoid catastrophe forgetting has not been well investigated.
1 code implementation • 27 Dec 2020 • Swapnil Parekh, Yaman Kumar Singla, Changyou Chen, Junyi Jessy Li, Rajiv Ratn Shah
However, little research has been put to understand and interpret the black-box nature of these deep-learning based scoring models.
no code implementations • 2 Dec 2020 • Yang Zhao, Chunyuan Li, Ping Yu, Changyou Chen
Few-shot learning features the capability of generalizing from a few examples.
no code implementations • NeurIPS 2020 • Fan Yang, Alina Vereshchaka, Changyou Chen, Wen Dong
We demonstrate the performance of our algorithm through benchmarking with three state-of-the-art multi-agent imitation learning algorithms on several tasks, including solving a multi-agent traffic optimization problem in a real-world transportation network.
1 code implementation • CVPR 2021 • Yang Zhao, Changyou Chen
Instead of explicitly extracting the two codes and applying adaptive instance normalization to combine them, our latent EBM can implicitly learn to transport the source style code to the target style code while preserving the content code, an advantage over existing image translation methods.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Ruiyi Zhang, Changyou Chen, Xinyuan Zhang, Ke Bai, Lawrence Carin
In sequence-to-sequence models, classical optimal transport (OT) can be applied to semantically match generated sentences with target sentences.
no code implementations • ICLR 2021 • Kevin J Liang, Weituo Hao, Dinghan Shen, Yufan Zhou, Weizhu Chen, Changyou Chen, Lawrence Carin
Large-scale language models have recently demonstrated impressive empirical performance.
no code implementations • NeurIPS 2020 • Yufan Zhou, Changyou Chen, Jinhui Xu
Manifold learning is a fundamental problem in machine learning with numerous applications.
no code implementations • EMNLP 2020 • Bang An, Jie Lyu, Zhenyi Wang, Chunyuan Li, Changwei Hu, Fei Tan, Ruiyi Zhang, Yifan Hu, Changyou Chen
The neural attention mechanism plays an important role in many natural language processing applications.
1 code implementation • ECCV 2020 • Ping Yu, Yang Zhao, Chunyuan Li, Junsong Yuan, Changyou Chen
Generating long-range skeleton-based human actions has been a challenging problem since small deviations of one frame can cause a malformed action sequence.
Ranked #2 on Human action generation on NTU RGB+D 2D
no code implementations • ACL 2020 • Lin Zheng, Qinliang Su, Dinghan Shen, Changyou Chen
Generative semantic hashing is a promising technique for large-scale information retrieval thanks to its fast retrieval speed and small memory footprint.
no code implementations • 16 May 2020 • Yufan Zhou, Jiayi Xian, Changyou Chen, Jinhui Xu
We then propose feature aggregation as the composition of the original neighbor-based kernel and a learnable kernel to encode feature similarities in a feature space.
no code implementations • 14 May 2020 • Tianhang Zheng, Sheng Liu, Changyou Chen, Junsong Yuan, Baochun Li, Kui Ren
We first formulate generation of adversarial skeleton actions as a constrained optimization problem by representing or approximating the physiological and physical constraints with mathematical formulations.
no code implementations • 4 May 2020 • Ruiyi Zhang, Tong Yu, Yilin Shen, Hongxia Jin, Changyou Chen, Lawrence Carin
Text-based interactive recommendation provides richer user feedback and has demonstrated advantages over traditional interactive recommender systems.
no code implementations • ACL 2020 • Ruiyi Zhang, Changyou Chen, Zhe Gan, Wenlin Wang, Dinghan Shen, Guoyin Wang, Zheng Wen, Lawrence Carin
Auto-regressive text generation models usually focus on local fluency, and may cause inconsistent semantic meaning in long text generation.
no code implementations • ACL 2020 • Zhenyi Wang, Xiaoyang Wang, Bang An, Dong Yu, Changyou Chen
Text generation from a knowledge base aims to translate knowledge triples to natural language descriptions.
1 code implementation • ICLR 2020 • Zhenyi Wang, Yang Zhao, Ping Yu, Ruiyi Zhang, Changyou Chen
Specifically, we propose a Bayesian meta sampling framework consisting of two main components: a meta sampler and a sample adapter.
no code implementations • 22 Apr 2020 • Yang Zhao, Ping Yu, Suchismit Mahapatra, Qinliang Su, Changyou Chen
Variational autoencoders (VAEs) are essential tools in end-to-end representation learning.
no code implementations • 21 Apr 2020 • Alexander Hanbo Li, Yaqing Wang, Changyou Chen, Jing Gao
Effective inference for a generative adversarial model remains an important and challenging problem.
2 code implementations • ICML 2020 • Yang Zhao, Chunyuan Li, Ping Yu, Jianfeng Gao, Changyou Chen
The instability in GAN training has been a long-standing problem despite remarkable research efforts.
Ranked #1 on Image-to-Image Translation on anime-to-selfie
no code implementations • 20 Jan 2020 • Ruiyi Zhang, Changyou Chen, Zhe Gan, Zheng Wen, Wenlin Wang, Lawrence Carin
Reinforcement learning (RL) has been widely studied for improving sequence-generation models.
1 code implementation • AAAI 2019 • Zhenyi Wang, Ping Yu, Yang Zhao, Ruiyi Zhang, Yufan Zhou, Junsong Yuan, Changyou Chen
In this paper, we focus on skeleton-based action generation and propose to model smooth and diverse transitions on a latent space of action sequences with much lower dimensionality.
Ranked #4 on Human action generation on NTU RGB+D 2D
no code implementations • 2 Dec 2019 • Yufan Zhou, Changyou Chen, Jinhui Xu
Learning with kernels is an important concept in machine learning.
no code implementations • NeurIPS 2019 • Ruiyi Zhang, Tong Yu, Yilin Shen, Hongxia Jin, Changyou Chen
Text-based interactive recommendation provides richer user preferences and has demonstrated advantages over traditional interactive recommender systems.
1 code implementation • 25 Nov 2019 • Yitong Yan, Chuangchuang Liu, Changyou Chen, Xianfang Sun, Longcun Jin, Xiang Zhou
Firstly, instead of producing a single score to discriminate images between real and fake, we propose a variant, called Fine-grained Attention Generative Adversarial Network for image super-resolution (FASRGAN), to discriminate each pixel between real and fake.
1 code implementation • IJCNLP 2019 • Le Fang, Chunyuan Li, Jianfeng Gao, Wen Dong, Changyou Chen
Deep latent variable models (LVM) such as variational auto-encoder (VAE) have recently played an important role in text generation.
no code implementations • IJCNLP 2019 • Wei Dong, Qinliang Su, Dinghan Shen, Changyou Chen
Hashing is promising for large-scale information retrieval tasks thanks to the efficiency of distance evaluation between binary codes.
no code implementations • 24 Jun 2019 • Jun Wen, Nenggan Zheng, Junsong Yuan, Zhefeng Gong, Changyou Chen
By imposing distribution matching on both features and labels (via uncertainty), label distribution mismatching in source and target data is effectively alleviated, encouraging the classifier to produce consistent predictions across domains.
no code implementations • NAACL 2019 • Wenlin Wang, Zhe Gan, Hongteng Xu, Ruiyi Zhang, Guoyin Wang, Dinghan Shen, Changyou Chen, Lawrence Carin
We propose a topic-guided variational auto-encoder (TGVAE) model for text generation.
no code implementations • 15 May 2019 • Bai Li, Changyou Chen, Wenlin Wang, Lawrence Carin
Adversarial examples are carefully perturbed in-puts for fooling machine learning models.
no code implementations • ICLR 2019 • Bai Li, Changyou Chen, Wenlin Wang, Lawrence Carin
In this paper, we propose a powerful second-order attack method that reduces the accuracy of the defense model by Madry et al. (2017).
no code implementations • 17 Mar 2019 • Wenlin Wang, Zhe Gan, Hongteng Xu, Ruiyi Zhang, Guoyin Wang, Dinghan Shen, Changyou Chen, Lawrence Carin
We propose a topic-guided variational autoencoder (TGVAE) model for text generation.
no code implementations • 19 Feb 2019 • Ruiyi Zhang, Zheng Wen, Changyou Chen, Lawrence Carin
Thompson sampling (TS) is a class of algorithms for sequential decision-making, which requires maintaining a posterior distribution over a model.
4 code implementations • ICLR 2020 • Ruqi Zhang, Chunyuan Li, Jianyi Zhang, Changyou Chen, Andrew Gordon Wilson
The posteriors over neural network weights are high dimensional and multimodal.
no code implementations • ICLR 2019 • Liqun Chen, Yizhe Zhang, Ruiyi Zhang, Chenyang Tao, Zhe Gan, Haichao Zhang, Bai Li, Dinghan Shen, Changyou Chen, Lawrence Carin
Sequence-to-sequence models are commonly trained via maximum likelihood estimation (MLE).
3 code implementations • 3 Jan 2019 • Chunyuan Li, Ke Bai, Jianqiao Li, Guoyin Wang, Changyou Chen, Lawrence Carin
We investigate adversarial learning in the case when only an unnormalized form of the density can be accessed, rather than samples.
3 code implementations • ICCV 2019 • Tianhang Zheng, Changyou Chen, Junsong Yuan, Bo Li, Kui Ren
Our motivation for constructing a saliency map is by point dropping, which is a non-differentiable operator.
no code implementations • 21 Nov 2018 • Yang Zhao, Jianyi Zhang, Changyou Chen
Scalable Bayesian sampling is playing an important role in modern machine learning, especially in the fast-developed unsupervised-(deep)-learning models.
no code implementations • ICML 2020 • Jianyi Zhang, Yang Zhao, Changyou Chen
Stochastic particle-optimization sampling (SPOS) is a recently-developed scalable Bayesian sampling framework that unifies stochastic gradient MCMC (SG-MCMC) and Stein variational gradient descent (SVGD) algorithms based on Wasserstein gradient flows.
no code implementations • 2 Nov 2018 • Ruiyi Zhang, Changyou Chen, Zhe Gan, Wenlin Wang, Liqun Chen, Dinghan Shen, Guoyin Wang, Lawrence Carin
Sequence generation with reinforcement learning (RL) has received significant attention recently.
no code implementations • 10 Oct 2018 • Tianhang Zheng, Changyou Chen, Kui Ren
In this paper, we give a negative answer by proposing a training paradigm that is comparable to PGD adversarial training on several standard datasets, while only using noisy-natural samples.
no code implementations • 27 Sep 2018 • Jianyi Zhang, Ruiyi Zhang, Changyou Chen
With such theoretical guarantees, SPOS can be safely and effectively applied on both Bayesian DL and deep RL tasks.
no code implementations • 27 Sep 2018 • Hanbo Li, Yaqing Wang, Changyou Chen, Jing Gao
We propose a novel approach, Adversarial Inference by Matching priors and conditionals (AIM), which explicitly matches prior and conditional distributions in both data and code spaces, and puts a direct constraint on the dependency structure of the generative model.
3 code implementations • NeurIPS 2019 • Bai Li, Changyou Chen, Wenlin Wang, Lawrence Carin
The existence of adversarial data examples has drawn significant attention in the deep-learning community; such data are seemingly minimally perturbed relative to the original data, but lead to very different outputs from a deep-learning algorithm.
no code implementations • 5 Sep 2018 • Jianyi Zhang, Ruiyi Zhang, Lawrence Carin, Changyou Chen
Particle-optimization-based sampling (POS) is a recently developed effective sampling technique that interactively updates a set of particles.
4 code implementations • 16 Aug 2018 • Tianhang Zheng, Changyou Chen, Kui Ren
Recent work on adversarial attack has shown that Projected Gradient Descent (PGD) Adversary is a universal first-order adversary, and the classifier adversarially trained by PGD is robust against a wide range of first-order attacks.
no code implementations • ICML 2018 • Ruiyi Zhang, Changyou Chen, Chunyuan Li, Lawrence Carin
Policy optimization is a core component of reinforcement learning (RL), and most existing RL methods directly optimize parameters of a policy based on maximizing the expected total reward, or its surrogate.
no code implementations • 29 May 2018 • Changyou Chen, Ruiyi Zhang, Wenlin Wang, Bai Li, Liqun Chen
There has been recent interest in developing scalable Bayesian sampling methods such as stochastic gradient MCMC (SG-MCMC) and Stein variational gradient descent (SVGD) for big-data analysis.
1 code implementation • 30 Dec 2017 • Ruiyi Zhang, Chunyuan Li, Changyou Chen, Lawrence Carin
Learning probability distributions on the weights of neural networks (NNs) has recently proven beneficial in many applications.
no code implementations • 25 Dec 2017 • Bai Li, Changyou Chen, Hao liu, Lawrence Carin
Significant success has been realized recently on applying machine learning to real-world applications.
no code implementations • 29 Nov 2017 • Changyou Chen, Ruiyi Zhang
Stochastic gradient Markov chain Monte Carlo (SG-MCMC) has been increasingly popular in Bayesian learning due to its ability to deal with large data.
no code implementations • 15 Nov 2017 • Wenlin Wang, Yunchen Pu, Vinay Kumar Verma, Kai Fan, Yizhe Zhang, Changyou Chen, Piyush Rai, Lawrence Carin
We present a deep generative model for learning to predict classes not seen at training time.
5 code implementations • NeurIPS 2017 • Chunyuan Li, Hao liu, Changyou Chen, Yunchen Pu, Liqun Chen, Ricardo Henao, Lawrence Carin
We investigate the non-identifiability issues associated with bidirectional adversarial training for joint distribution matching.
no code implementations • ICML 2018 • Changyou Chen, Chunyuan Li, Liqun Chen, Wenlin Wang, Yunchen Pu, Lawrence Carin
Distinct from normalizing flows and GANs, CTFs can be adopted to achieve the above two goals in one framework, with theoretical guarantees.
no code implementations • 4 Sep 2017 • Changyou Chen, Wenlin Wang, Yizhe Zhang, Qinliang Su, Lawrence Carin
However, there has been little theoretical analysis of the impact of minibatch size to the algorithm's convergence rate.
no code implementations • ICML 2017 • Yizhe Zhang, Changyou Chen, Zhe Gan, Ricardo Henao, Lawrence Carin
A framework is proposed to improve the sampling efficiency of stochastic gradient MCMC, based on Hamiltonian Monte Carlo.
no code implementations • ACL 2017 • Zhe Gan, Chunyuan Li, Changyou Chen, Yunchen Pu, Qinliang Su, Lawrence Carin
Recurrent neural networks (RNNs) have shown promising performance for language modeling.
no code implementations • 14 Nov 2016 • Wenlin Wang, Changyou Chen, Wenqi Wang, Piyush Rai, Lawrence Carin
Unlike most existing methods for early classification of time series data, that are designed to solve this problem under the assumption of the availability of a good set of pre-defined (often hand-crafted) features, our framework can jointly perform feature learning (by learning a deep hierarchy of \emph{shapelets} capturing the salient characteristics in each time series), along with a dynamic truncation model to help our deep feature learning architecture focus on the early parts of each time series.
no code implementations • NeurIPS 2015 • Changyou Chen, Nan Ding, Lawrence Carin
Our theoretical results show faster convergence rates and more accurate invariant measures for SG-MCMCs with higher-order integrators.
no code implementations • NeurIPS 2016 • Changyou Chen, Nan Ding, Chunyuan Li, Yizhe Zhang, Lawrence Carin
In this paper we develop theory to show that while the bias and MSE of an SG-MCMC algorithm depend on the staleness of stochastic gradients, its estimation variance (relative to the expected estimate, based on a prescribed number of samples) is independent of it.
no code implementations • 22 Sep 2016 • Kar Wai Lim, Wray Buntine, Changyou Chen, Lan Du
In this article, we present efficient methods for the use of these processes in this hierarchical context, and apply them to latent variable models for text analytics.
no code implementations • 22 Sep 2016 • Kar Wai Lim, Changyou Chen, Wray Buntine
Exploiting this additional information, we propose the Twitter-Network (TN) topic model to jointly model the text and the social network in a full Bayesian nonparametric way.
no code implementations • 2 Jun 2016 • Qinliang Su, Xuejun Liao, Changyou Chen, Lawrence Carin
We introduce the truncated Gaussian graphical model (TGGM) as a novel framework for designing statistical models for nonlinear learning.
no code implementations • CVPR 2016 • Chunyuan Li, Andrew Stevens, Changyou Chen, Yunchen Pu, Zhe Gan, Lawrence Carin
Learning the representation of shape cues in 2D & 3D objects for recognition is a fundamental task in computer vision.
no code implementations • NeurIPS 2016 • Yizhe Zhang, Xiangyu Wang, Changyou Chen, Ricardo Henao, Kai Fan, Lawrence Carin
We unify slice sampling and Hamiltonian Monte Carlo (HMC) sampling, demonstrating their connection via the Hamiltonian-Jacobi equation from Hamiltonian mechanics.
1 code implementation • 25 Dec 2015 • Changyou Chen, David Carlson, Zhe Gan, Chunyuan Li, Lawrence Carin
Stochastic gradient Markov chain Monte Carlo (SG-MCMC) methods are Bayesian analogs to popular stochastic optimization methods; however, this connection is not well studied.
no code implementations • 23 Dec 2015 • Chunyuan Li, Changyou Chen, David Carlson, Lawrence Carin
Pytorch implementations of Bayes By Backprop, MC Dropout, SGLD, the Local Reparametrization Trick, KF-Laplace and more
no code implementations • 23 Dec 2015 • Chunyuan Li, Changyou Chen, Kai Fan, Lawrence Carin
Stochastic gradient MCMC algorithms (SG-MCMC) are a family of diffusion-based sampling methods for large-scale Bayesian learning.
no code implementations • 18 Aug 2015 • Changwei Hu, Piyush Rai, Changyou Chen, Matthew Harding, Lawrence Carin
We present a Bayesian non-negative tensor factorization model for count-valued tensor data, and develop scalable inference algorithms (both batch and online) for dealing with massive tensors.
no code implementations • NeurIPS 2014 • Nan Ding, Youhan Fang, Ryan Babbush, Changyou Chen, Robert D. Skeel, Hartmut Neven
To remedy this problem, we show that one can leverage a small number of additional variables in order to stabilize momentum fluctuations induced by the unknown noise.
no code implementations • NeurIPS 2014 • Changyou Chen, Jun Zhu, Xinhua Zhang
We present max-margin Bayesian clustering (BMC), a general and robust framework that incorporates the max-margin criterion into Bayesian clustering models, as well as two concrete models of BMC to demonstrate its flexibility and effectiveness in dealing with different clustering tasks.