no code implementations • 2 Nov 2022 • Chenlin Meng, Kristy Choi, Jiaming Song, Stefano Ermon
To this end, we propose an analogous score function called the "Concrete score", a generalization of the (Stein) score for discrete settings.
no code implementations • 22 Oct 2022 • Kristy Choi, Chris Cundy, Sanjari Srivastava, Stefano Ermon
Particularly in low-data regimes, an outstanding challenge in machine learning is developing principled techniques for augmenting our models with suitable priors.
no code implementations • 28 Sep 2022 • Chenlin Meng, Linqi Zhou, Kristy Choi, Tri Dao, Stefano Ermon
Normalizing flows model complex probability distributions using maps obtained by composing invertible layers.
no code implementations • 4 Jul 2022 • Saeid Asgari Taghanaki, Ali Gholami, Fereshte Khani, Kristy Choi, Linh Tran, Ran Zhang, Aliasghar Khani
Batch normalization (BN) is a ubiquitous technique for training deep neural networks that accelerates their convergence to reach higher accuracy.
1 code implementation • 22 Nov 2021 • Kristy Choi, Chenlin Meng, Yang song, Stefano Ermon
We then estimate the instantaneous rate of change of the bridge distributions indexed by time (the "time score") -- a quantity defined analogously to data (Stein) scores -- with a novel time score matching objective.
1 code implementation • 5 Jul 2021 • Kristy Choi, Madeline Liao, Stefano Ermon
Density ratio estimation serves as an important technique in the unsupervised machine learning toolbox.
no code implementations • 11 Jun 2021 • Saeid Asgari Taghanaki, Kristy Choi, Amir Khasahmadi, Anirudh Goyal
A fundamental challenge in artificial intelligence is learning useful representations of data that yield good performance on a downstream task, without overfitting to spurious input features.
no code implementations • 15 Feb 2021 • Berivan Isik, Kristy Choi, Xin Zheng, Tsachy Weissman, Stefano Ermon, H. -S. Philip Wong, Armin Alaghi
Compression and efficient storage of neural network (NN) parameters is critical for applications that run on resource-constrained devices.
no code implementations • 1 Jan 2021 • Saeid Asgari, Kristy Choi, Amir Hosein Khasahmadi, Anirudh Goyal
A fundamental challenge in artificial intelligence is learning useful representations of data that yield good performance on a downstream classification task, without overfitting to spurious input features.
no code implementations • NeurIPS Workshop DL-IG 2020 • Berivan Isik, Kristy Choi, Xin Zheng, H.-S. Philip Wong, Stefano Ermon, Tsachy Weissman, Armin Alaghi
Efficient compression and storage of neural network (NN) parameters is critical for resource-constrained, downstream machine learning applications.
no code implementations • ICML 2020 • Kristy Choi, Curtis Hawthorne, Ian Simon, Monica Dinculescu, Jesse Engel
We consider the problem of learning high-level controls over the global structure of generated sequences, particularly in the context of symbolic music generation with complex language models.
1 code implementation • ICML 2020 • Kristy Choi, Aditya Grover, Trisha Singh, Rui Shu, Stefano Ermon
Real-world datasets are often biased with respect to key demographic factors such as race and gender.
1 code implementation • 5 Feb 2019 • Mike Wu, Kristy Choi, Noah Goodman, Stefano Ermon
Despite the recent success in probabilistic modeling and their applications, generative models trained using traditional inference techniques struggle to adapt to new distributions, even when the target distribution may be closely related to the ones seen during training.
1 code implementation • 19 Nov 2018 • Kristy Choi, Kedar Tatwawadi, Aditya Grover, Tsachy Weissman, Stefano Ermon
For reliable transmission across a noisy communication channel, classical results from information theory show that it is asymptotically optimal to separate out the source and channel coding processes.
no code implementations • 27 Sep 2018 • Kristy Choi, Kedar Tatwawadi, Tsachy Weissman, Stefano Ermon
For reliable transmission across a noisy communication channel, classical results from information theory show that it is asymptotically optimal to separate out the source and channel coding processes.