Search Results for author: Chongyang Gao

Found 17 papers, 12 papers with code

Higher Layers Need More LoRA Experts

1 code implementation13 Feb 2024 Chongyang Gao, Kezhen Chen, Jinmeng Rao, Baochen Sun, Ruibo Liu, Daiyi Peng, Yawen Zhang, Xiaoyuan Guo, Jie Yang, VS Subrahmanian

In this paper, we introduce a novel parameter-efficient MoE method, \textit{\textbf{M}oE-L\textbf{o}RA with \textbf{L}ayer-wise Expert \textbf{A}llocation (MoLA)} for Transformer-based models, where each model layer has the flexibility to employ a varying number of LoRA experts.

How to Configure Good In-Context Sequence for Visual Question Answering

1 code implementation4 Dec 2023 Li Li, Jiawei Peng, Huiyi Chen, Chongyang Gao, Xu Yang

Inspired by the success of Large Language Models in dealing with new tasks via In-Context Learning (ICL) in NLP, researchers have also developed Large Vision-Language Models (LVLMs) with ICL capabilities.

In-Context Learning Question Answering +2

Improving Representation Learning for Histopathologic Images with Cluster Constraints

1 code implementation ICCV 2023 Weiyi Wu, Chongyang Gao, Joseph DiPalma, Soroush Vosoughi, Saeed Hassanpour

This framework aims for transferable representation learning and semantically meaningful clustering by synergizing invariance loss and clustering loss in WSI analysis.

Clustering Representation Learning +1

Bootstrapping Vision-Language Learning with Decoupled Language Pre-training

1 code implementation NeurIPS 2023 Yiren Jian, Chongyang Gao, Soroush Vosoughi

We present a novel methodology aimed at optimizing the application of frozen large language models (LLMs) for resource-intensive vision-language (VL) pre-training.

Joint Latent Topic Discovery and Expectation Modeling for Financial Markets

no code implementations1 Jun 2023 Lili Wang, Chenghan Huang, Chongyang Gao, Weicheng Ma, Soroush Vosoughi

In the pursuit of accurate and scalable quantitative methods for financial market analysis, the focus has shifted from individual stock models to those capturing interrelations between companies and their stocks.

Knowledge from Large-Scale Protein Contact Prediction Models Can Be Transferred to the Data-Scarce RNA Contact Prediction Task

1 code implementation13 Feb 2023 Yiren Jian, Chongyang Gao, Chen Zeng, Yunjie Zhao, Soroush Vosoughi

Our findings indicate that the learned structural patterns of proteins can be transferred to RNAs, opening up potential new avenues for research.

Transfer Learning

Learning to Collocate Visual-Linguistic Neural Modules for Image Captioning

1 code implementation4 Oct 2022 Xu Yang, Hanwang Zhang, Chongyang Gao, Jianfei Cai

This is because the language is only partially observable, for which we need to dynamically collocate the modules during the process of image captioning.

Image Captioning Sentence +2

Non-Linguistic Supervision for Contrastive Learning of Sentence Embeddings

1 code implementation20 Sep 2022 Yiren Jian, Chongyang Gao, Soroush Vosoughi

This indicates that Transformer models are able to generalize better by doing a similar task (i. e., clustering) with unpaired examples from different modalities in a multi-task fashion.

Clustering Contrastive Learning +3

Embedding Hallucination for Few-Shot Language Fine-tuning

1 code implementation NAACL 2022 Yiren Jian, Chongyang Gao, Soroush Vosoughi

Few-shot language learners adapt knowledge from a pre-trained model to recognize novel classes from a few-labeled sentences.

Data Augmentation Hallucination +1

Contrastive Learning for Prompt-Based Few-Shot Language Learners

1 code implementation NAACL 2022 Yiren Jian, Chongyang Gao, Soroush Vosoughi

Following this line of work, we present a contrastive learning framework that clusters inputs from the same class for better generality of models trained with only limited examples.

Contrastive Learning In-Context Learning +2

Non-Parallel Text Style Transfer with Self-Parallel Supervision

1 code implementation ICLR 2022 Ruibo Liu, Chongyang Gao, Chenyan Jia, Guangxuan Xu, Soroush Vosoughi

The performance of existing text style transfer models is severely limited by the non-parallel datasets on which the models are trained.

Imitation Learning Style Transfer +1

Knowledge Infused Decoding

1 code implementation ICLR 2022 Ruibo Liu, Guoqing Zheng, Shashank Gupta, Radhika Gaonkar, Chongyang Gao, Soroush Vosoughi, Milad Shokouhi, Ahmed Hassan Awadallah

Hence, they tend to suffer from counterfactual or hallucinatory generation when used in knowledge-intensive natural language generation (NLG) tasks.

counterfactual Question Answering +1

MetaPix: Domain Transfer for Semantic Segmentation by Meta Pixel Weighting

1 code implementation5 Oct 2021 Yiren Jian, Chongyang Gao

Previous work has shown that the performance of a semantic segmentation model can be improved by training jointly with real and synthetic examples with a proper weighting on the synthetic data.

Meta-Learning Segmentation +1

Auto-Parsing Network for Image Captioning and Visual Question Answering

no code implementations ICCV 2021 Xu Yang, Chongyang Gao, Hanwang Zhang, Jianfei Cai

We propose an Auto-Parsing Network (APN) to discover and exploit the input data's hidden tree structures for improving the effectiveness of the Transformer-based vision-language systems.

Image Captioning Question Answering +1

Hierarchical Scene Graph Encoder-Decoder for Image Paragraph Captioning

no code implementations ACM International Conference on Multimedia 2020 Yang, Xu, Chongyang Gao, Hanwang Zhang, and Jianfei Cai

We propose irredundant attention in SSG-RNN to improve the possibility of abstracting topics from rarely described sub-graphs and inheriting attention in WSG-RNN to generate more grounded sentences with the abstracted topics, both of which give rise to more distinctive paragraphs.

Image Paragraph Captioning Sentence

Cannot find the paper you are looking for? You can Submit a new open access paper.