1 code implementation • 13 Jan 2025 • Xingchen Liu, Piyush Tayal, Jianyuan Wang, Jesus Zarzar, Tom Monnier, Konstantinos Tertikas, Jiali Duan, Antoine Toisoul, Jason Y. Zhang, Natalia Neverova, Andrea Vedaldi, Roman Shapovalov, David Novotny
We introduce Uncommon Objects in 3D (uCO3D), a new object-centric dataset for 3D deep learning and 3D generative AI.
no code implementations • 4 Aug 2022 • Yilei Zeng, Jiali Duan, Yang Li, Emilio Ferrara, Lerrel Pinto, C. -C. Jay Kuo, Stefanos Nikolaidis
In this work, we guide the curriculum reinforcement learning results towards a preferred performance level that is neither too hard nor too easy via learning from the human decision process.
no code implementations • 31 Jul 2022 • Xiaoyuan Guo, Jiali Duan, C. -C. Jay Kuo, Judy Wawira Gichoya, Imon Banerjee
Language modality within the vision language pretraining framework is innately discretized, endowing each word in the language vocabulary a semantic meaning.
no code implementations • 6 Jun 2022 • Dongsheng Ding, Kaiqing Zhang, Jiali Duan, Tamer Başar, Mihailo R. Jovanović
We study sequential decision making problems aimed at maximizing the expected total reward while satisfying a constraint on the expected total utility.
no code implementations • 6 Apr 2022 • Xiaoyuan Guo, Jiali Duan, Saptarshi Purkayastha, Hari Trivedi, Judy Wawira Gichoya, Imon Banerjee
While existing methods can be applied for class-wise retrieval (aka.
no code implementations • CVPR 2022 • Jiali Duan, Liqun Chen, Son Tran, Jinyu Yang, Yi Xu, Belinda Zeng, Trishul Chilimbi
Aligning signals from different modalities is an important step in vision-language representation learning as it affects the performance of later stages such as cross-modality fusion.
1 code implementation • CVPR 2022 • Jinyu Yang, Jiali Duan, Son Tran, Yi Xu, Sampath Chanda, Liqun Chen, Belinda Zeng, Trishul Chilimbi, Junzhou Huang
Besides CMA, TCL introduces an intra-modal contrastive objective to provide complementary benefits in representation learning.
Ranked #4 on Zero-Shot Cross-Modal Retrieval on COCO 2014
no code implementations • 29 Jul 2021 • Jiali Duan, C. -C. Jay Kuo
The fact that there exists a gap between low-level features and semantic meanings of images, called the semantic gap, is known for decades.
no code implementations • CVPR 2021 • Jiali Duan, Yen-Liang Lin, Son Tran, Larry S. Davis, C. -C. Jay Kuo
We first train a teacher model on the labeled data and use it to generate pseudo labels for the unlabeled data.
no code implementations • 11 Nov 2018 • Yao Zhu, Saksham Suri, Pranav Kulkarni, Yueru Chen, Jiali Duan, C. -C. Jay Kuo
An interpretable generative model for handwritten digits synthesis is proposed in this work.
2 code implementations • 5 Oct 2018 • C. -C. Jay Kuo, Min Zhang, Siyang Li, Jiali Duan, Yueru Chen
To construct convolutional layers, we develop a new signal transform, called the Saab (Subspace Approximation with Adjusted Bias) transform.
no code implementations • 5 Jul 2018 • Jiali Duan, Xiaoyuan Guo, Yuhang Song, Chao Yang, C. -C. Jay Kuo
Previous methods have dealt with discrete manipulation of facial attributes such as smile, sad, angry, surprise etc, out of canonical expressions and they are not scalable, operating in single modality.
no code implementations • 21 Nov 2016 • Jiali Duan, Shuai Zhou, Jun Wan, Xiaoyuan Guo, Stan Z. Li
Recently, the popularity of depth-sensors such as Kinect has made depth videos easily available while its advantages have not been fully exploited.