Search Results for author: Anil Kag

Found 14 papers, 9 papers with code

TextCraftor: Your Text Encoder Can be Image Quality Controller

no code implementations27 Mar 2024 Yanyu Li, Xian Liu, Anil Kag, Ju Hu, Yerlan Idelbayev, Dhritiman Sagar, Yanzhi Wang, Sergey Tulyakov, Jian Ren

Our findings reveal that, instead of replacing the CLIP text encoder used in Stable Diffusion with other large language models, we can enhance it through our proposed fine-tuning approach, TextCraftor, leading to substantial improvements in quantitative benchmarks and human assessments.

Image Generation

Snap Video: Scaled Spatiotemporal Transformers for Text-to-Video Synthesis

no code implementations22 Feb 2024 Willi Menapace, Aliaksandr Siarohin, Ivan Skorokhodov, Ekaterina Deyneka, Tsai-Shien Chen, Anil Kag, Yuwei Fang, Aleksei Stoliar, Elisa Ricci, Jian Ren, Sergey Tulyakov

Since video content is highly redundant, we argue that naively bringing advances of image models to the video generation domain reduces motion fidelity, visual quality and impairs scalability.

Image Generation Text-to-Video Generation +1

Scaffolding a Student to Instill Knowledge

1 code implementation International Conference on Learning Representations 2023 Anil Kag, Durmus Alp Emre Acar, Aditya Gangrade, Venkatesh Saligrama

We propose a novel knowledge distillation (KD) method to selectively instill teacher knowledge into a student model motivated by situations where the student's capacity is significantly smaller than that of the teachers.

Knowledge Distillation

Condensing CNNs With Partial Differential Equations

1 code implementation CVPR 2022 Anil Kag, Venkatesh Saligrama

Convolutional neural networks (CNNs) rely on the depth of the architecture to obtain complex features.

Hybrid Cloud-Edge Networks for Efficient Inference

1 code implementation29 Sep 2021 Anil Kag, Igor Fedorov, Aditya Gangrade, Paul Whatmough, Venkatesh Saligrama

The first network is a low-capacity network that can be deployed on an edge device, whereas the second is a high-capacity network deployed in the cloud.

Training Recurrent Neural Networks via Forward Propagation Through Time

1 code implementation International Conference on Machine Learning 2021 Anil Kag, Venkatesh Saligrama

BPTT updates RNN parameters on an instance by back-propagating the error in time over the entire sequence length, and as a result, leads to poor trainability due to the well-known gradient explosion/decay phenomena.

Time Adaptive Recurrent Neural Network

1 code implementation CVPR 2021 Anil Kag, Venkatesh Saligrama

We propose a learning method that, dynamically modifies the time-constants of the continuous-time counterpart of a vanilla RNN.

Selective Classification via One-Sided Prediction

1 code implementation15 Oct 2020 Aditya Gangrade, Anil Kag, Venkatesh Saligrama

We propose a novel method for selective classification (SC), a problem which allows a classifier to abstain from predicting some instances, thus trading off accuracy against coverage (the fraction of instances predicted).

Classification General Classification +1

RNNs Incrementally Evolving on an Equilibrium Manifold: A Panacea for Vanishing and Exploding Gradients?

1 code implementation ICLR 2020 Anil Kag, Ziming Zhang, Venkatesh Saligrama

Recurrent neural networks (RNNs) are particularly well-suited for modeling long-term dependencies in sequential data, but are notoriously hard to train because the error backpropagated in time either vanishes or explodes at an exponential rate.

RNNs Evolving on an Equilibrium Manifold: A Panacea for Vanishing and Exploding Gradients?

no code implementations22 Aug 2019 Anil Kag, Ziming Zhang, Venkatesh Saligrama

Recurrent neural networks (RNNs) are particularly well-suited for modeling long-term dependencies in sequential data, but are notoriously hard to train because the error backpropagated in time either vanishes or explodes at an exponential rate.

Equilibrated Recurrent Neural Network: Neuronal Time-Delayed Self-Feedback Improves Accuracy and Stability

no code implementations2 Mar 2019 Ziming Zhang, Anil Kag, Alan Sullivan, Venkatesh Saligrama

We show that such self-feedback helps stabilize the hidden state transitions leading to fast convergence during training while efficiently learning discriminative latent features that result in state-of-the-art results on several benchmark datasets at test-time.

Learning Compact Networks via Adaptive Network Regularization

no code implementations NIPS Workshop CDNNRIA 2018 Sivaramakrishnan Sankarapandian, Anil Kag, Rachel Manzelli, Brian Kulis

We describe a training strategy that grows the number of units during training, and show on several benchmark datasets that our model yields architectures that are smaller than those obtained when tuning the number of hidden units on a standard fixed architecture.

Cannot find the paper you are looking for? You can Submit a new open access paper.