Search Results for author: Junjie Ke

Found 12 papers, 4 papers with code

Optical Diffusion Models for Image Generation

no code implementations15 Jul 2024 Ilker Oguz, Niyazi Ulas Dinc, Mustafa Yıldırım, Junjie Ke, Innfarn Yoo, Qifei Wang, Feng Yang, Christophe Moser, Demetri Psaltis

Diffusion models generate new samples by progressively decreasing the noise from the initially provided random distribution.

Denoising Image Generation

Parrot: Pareto-optimal Multi-Reward Reinforcement Learning Framework for Text-to-Image Generation

no code implementations11 Jan 2024 Seung Hyun Lee, Yinxiao Li, Junjie Ke, Innfarn Yoo, Han Zhang, Jiahui Yu, Qifei Wang, Fei Deng, Glenn Entis, Junfeng He, Gang Li, Sangpil Kim, Irfan Essa, Feng Yang

We use the novel multi-reward optimization algorithm to jointly optimize the T2I model and a prompt expansion network, resulting in significant improvement of image quality and also allow to control the trade-off of different rewards using a reward related prompt during inference.

Reinforcement Learning (RL) Text-to-Image Generation

Rich Human Feedback for Text-to-Image Generation

1 code implementation CVPR 2024 Youwei Liang, Junfeng He, Gang Li, Peizhao Li, Arseniy Klimovskiy, Nicholas Carolan, Jiao Sun, Jordi Pont-Tuset, Sarah Young, Feng Yang, Junjie Ke, Krishnamurthy Dj Dvijotham, Katie Collins, Yiwen Luo, Yang Li, Kai J Kohlhoff, Deepak Ramachandran, Vidhya Navalpakkam

We show that the predicted rich human feedback can be leveraged to improve image generation, for example, by selecting high-quality training data to finetune and improve the generative models, or by creating masks with predicted heatmaps to inpaint the problematic regions.

Text-to-Image Generation

Forward-Forward Training of an Optical Neural Network

no code implementations30 May 2023 Ilker Oguz, Junjie Ke, Qifei Wang, Feng Yang, Mustafa Yıldırım, Niyazi Ulas Dinc, Jih-Liang Hsieh, Christophe Moser, Demetri Psaltis

Neural networks (NN) have demonstrated remarkable capabilities in various tasks, but their computation-intensive nature demands faster and more energy-efficient hardware implementations.

VILA: Learning Image Aesthetics from User Comments with Vision-Language Pretraining

2 code implementations CVPR 2023 Junjie Ke, Keren Ye, Jiahui Yu, Yonghui Wu, Peyman Milanfar, Feng Yang

Our results show that our pretrained aesthetic vision-language model outperforms prior works on image aesthetic captioning over the AVA-Captions dataset, and it has powerful zero-shot capability for aesthetic tasks such as zero-shot style classification and zero-shot IAA, surpassing many supervised baselines.

Decoder Language Modelling +1

MRET: Multi-resolution Transformer for Video Quality Assessment

no code implementations13 Mar 2023 Junjie Ke, Tianhao Zhang, Yilin Wang, Peyman Milanfar, Feng Yang

No-reference video quality assessment (NR-VQA) for user generated content (UGC) is crucial for understanding and improving visual experience.

Video Quality Assessment Video Recognition +1

Rich Features for Perceptual Quality Assessment of UGC Videos

no code implementations CVPR 2021 Yilin Wang, Junjie Ke, Hossein Talebi, Joong Gon Yim, Neil Birkbeck, Balu Adsumilli, Peyman Milanfar, Feng Yang

Besides the subjective ratings and content labels of the dataset, we also propose a DNN-based framework to thoroughly analyze importance of content, technical quality, and compression level in perceptual quality.

Video Quality Assessment

Multi-path Neural Networks for On-device Multi-domain Visual Classification

no code implementations10 Oct 2020 Qifei Wang, Junjie Ke, Joshua Greaves, Grace Chu, Gabriel Bender, Luciano Sbaiz, Alec Go, Andrew Howard, Feng Yang, Ming-Hsuan Yang, Jeff Gilbert, Peyman Milanfar

This approach effectively reduces the total number of parameters and FLOPS, encouraging positive knowledge transfer while mitigating negative interference across domains.

General Classification Neural Architecture Search +1

Cannot find the paper you are looking for? You can Submit a new open access paper.