Search Results for author: Junjie Ke

Found 9 papers, 3 papers with code

Parrot: Pareto-optimal Multi-Reward Reinforcement Learning Framework for Text-to-Image Generation

no code implementations11 Jan 2024 Seung Hyun Lee, Yinxiao Li, Junjie Ke, Innfarn Yoo, Han Zhang, Jiahui Yu, Qifei Wang, Fei Deng, Glenn Entis, Junfeng He, Gang Li, Sangpil Kim, Irfan Essa, Feng Yang

Additionally, Parrot employs a joint optimization approach for the T2I model and the prompt expansion network, facilitating the generation of quality-aware text prompts, thus further enhancing the final image quality.

Reinforcement Learning (RL) Text-to-Image Generation

Rich Human Feedback for Text-to-Image Generation

1 code implementation15 Dec 2023 Youwei Liang, Junfeng He, Gang Li, Peizhao Li, Arseniy Klimovskiy, Nicholas Carolan, Jiao Sun, Jordi Pont-Tuset, Sarah Young, Feng Yang, Junjie Ke, Krishnamurthy Dj Dvijotham, Katie Collins, Yiwen Luo, Yang Li, Kai J Kohlhoff, Deepak Ramachandran, Vidhya Navalpakkam

We show that the predicted rich human feedback can be leveraged to improve image generation, for example, by selecting high-quality training data to finetune and improve the generative models, or by creating masks with predicted heatmaps to inpaint the problematic regions.

Text-to-Image Generation

Forward-Forward Training of an Optical Neural Network

no code implementations30 May 2023 Ilker Oguz, Junjie Ke, Qifei Wang, Feng Yang, Mustafa Yıldırım, Niyazi Ulas Dinc, Jih-Liang Hsieh, Christophe Moser, Demetri Psaltis

Neural networks (NN) have demonstrated remarkable capabilities in various tasks, but their computation-intensive nature demands faster and more energy-efficient hardware implementations.

VILA: Learning Image Aesthetics from User Comments with Vision-Language Pretraining

1 code implementation CVPR 2023 Junjie Ke, Keren Ye, Jiahui Yu, Yonghui Wu, Peyman Milanfar, Feng Yang

Our results show that our pretrained aesthetic vision-language model outperforms prior works on image aesthetic captioning over the AVA-Captions dataset, and it has powerful zero-shot capability for aesthetic tasks such as zero-shot style classification and zero-shot IAA, surpassing many supervised baselines.

Language Modelling Video Quality Assessment

MRET: Multi-resolution Transformer for Video Quality Assessment

no code implementations13 Mar 2023 Junjie Ke, Tianhao Zhang, Yilin Wang, Peyman Milanfar, Feng Yang

No-reference video quality assessment (NR-VQA) for user generated content (UGC) is crucial for understanding and improving visual experience.

Video Quality Assessment Video Recognition +1

Rich Features for Perceptual Quality Assessment of UGC Videos

no code implementations CVPR 2021 Yilin Wang, Junjie Ke, Hossein Talebi, Joong Gon Yim, Neil Birkbeck, Balu Adsumilli, Peyman Milanfar, Feng Yang

Besides the subjective ratings and content labels of the dataset, we also propose a DNN-based framework to thoroughly analyze importance of content, technical quality, and compression level in perceptual quality.

Video Quality Assessment

Multi-path Neural Networks for On-device Multi-domain Visual Classification

no code implementations10 Oct 2020 Qifei Wang, Junjie Ke, Joshua Greaves, Grace Chu, Gabriel Bender, Luciano Sbaiz, Alec Go, Andrew Howard, Feng Yang, Ming-Hsuan Yang, Jeff Gilbert, Peyman Milanfar

This approach effectively reduces the total number of parameters and FLOPS, encouraging positive knowledge transfer while mitigating negative interference across domains.

General Classification Neural Architecture Search +1

Cannot find the paper you are looking for? You can Submit a new open access paper.