no code implementations • 2 Dec 2024 • Xiang Li, Yucheng Zhou, Laiping Zhao, Jing Li, Fangming Liu
Moreover, we propose a detection framework tailored to this problem, which employs context augmentation modeling and multi-round iterative training.
1 code implementation • 26 Nov 2024 • Guanjie Chen, Xinyu Zhao, Yucheng Zhou, Tianlong Chen, Yu Cheng
Diffusion Transformers (DiT), an emerging image and video generation model architecture, has demonstrated great potential because of its high generation quality and scalability properties.
no code implementations • 25 Oct 2024 • Yucheng Zhou, Zhi Rao, Jun Wan, Jianbing Shen
Large Vision-Language Models (LVLMs) excel in cross-model tasks but experience performance declines in long-context reasoning due to overreliance on textual information and reduced visual dependency.
1 code implementation • 21 Sep 2024 • Jiashuo Sun, Jihai Zhang, Yucheng Zhou, Zhaochen Su, Xiaoye Qu, Yu Cheng
To address these challenges, we propose a self-refinement framework designed to teach LVLMs to Selectively Utilize Retrieved Information (SURf).
no code implementations • 6 May 2024 • Qianning Wang, He Hu, Yucheng Zhou
Existing vision models for defect recognition methods are insufficient for handling the complexities and variations of defects in contemporary manufacturing settings.
no code implementations • 11 Apr 2024 • Handi Deng, Yucheng Zhou, Jiaxuan Xiang, Liujie Gu, Yan Luo, Hai Feng, Mingyuan Liu, Cheng Ma
Foundation models have rapidly evolved and have achieved significant accomplishments in computer vision tasks.
no code implementations • 4 Apr 2024 • Qianning Wang, Chenglin Wang, Zhixin Lai, Yucheng Zhou
The classification of insect pests is a critical task in agricultural technology, vital for ensuring food security and environmental sustainability.
1 code implementation • 26 Mar 2024 • Zhixin Lai, Jing Wu, Suiyao Chen, Yucheng Zhou, Naira Hovakimyan
In this study, we uncover the unexpected efficacy of residual-based large language models (LLMs) as part of encoders for biomedical imaging tasks, a domain traditionally devoid of language or textual data.
no code implementations • 26 Mar 2024 • Wei Tao, Yucheng Zhou, Yanlin Wang, Wenqiang Zhang, Hongyu Zhang, Yu Cheng
To overcome this challenge, we empirically study the reason why LLMs fail to resolve GitHub issues and analyze the major factors.
no code implementations • 18 Feb 2024 • Yucheng Zhou, Xiang Li, Qianning Wang, Jianbing Shen
In Large Visual Language Models (LVLMs), the efficacy of In-Context Learning (ICL) remains limited by challenges in cross-modal interactions and representation disparities.
1 code implementation • 16 Jan 2024 • Wei Tao, Yucheng Zhou, Yanlin Wang, Hongyu Zhang, Haofen Wang, Wenqiang Zhang
However, previous methods are trained on the entire dataset without considering the fact that a portion of commit messages adhere to good practice (i. e., good-practice commits), while the rest do not.
no code implementations • 15 Nov 2023 • Yucheng Zhou, Xiubo Geng, Tao Shen, Chongyang Tao, Guodong Long, Jian-Guang Lou, Jianbing Shen
Large Language Models (LLMs) have ushered in a transformative era in the field of natural language processing, excelling in tasks related to text comprehension and generation.
no code implementations • 20 Oct 2023 • Jaeseoung Park, Ashwani Kumar, Yucheng Zhou, Sangheon Oh, Jeong-Hoon Kim, Yuhan Shi, Soumil Jain, Gopabandhu Hota, Amelie L. Nagle, Catherine D. Schuman, Gert Cauwenberghs, Duygu Kuzum
To address all these challenges, we developed a forming-free and bulk switching RRAM technology based on a trilayer metal-oxide stack.
no code implementations • 26 Jan 2023 • Yucheng Zhou, Guodong Long
Existing multi-style image captioning methods show promising results in generating a caption with accurate visual content and desired linguistic style.
no code implementations • 26 Jan 2023 • Yucheng Zhou, Guodong Long
Specifically, we construct visual and semantic event graphs from story plots and ending image, and leverage event-based reasoning to reason and mine implicit information in a single modality.
no code implementations • 26 Jan 2023 • Yucheng Zhou, Guodong Long
Text-guided image inpainting (TGII) aims to restore missing regions based on a given text in a damaged image.
no code implementations • 20 Dec 2022 • Yucheng Zhou, Tao Shen, Xiubo Geng, Chongyang Tao, Guodong Long, Can Xu, Daxin Jiang
Long document retrieval aims to fetch query-relevant documents from a large-scale collection, where knowledge distillation has become de facto to improve a retriever by mimicking a heterogeneous yet powerful cross-encoder.
no code implementations • 27 Oct 2022 • Xiang Li, Yucheng Zhou
Researching bragging behavior on social media arouses interest of computational (socio) linguists.
no code implementations • 27 Oct 2022 • Chenglin Wang, Yucheng Zhou, Guodong Long, Xiaodong Wang, Xiaowei Xu
Therefore, we propose an unsupervised knowledge graph construction method to build a scientific knowledge graph (SKG) without any labeled data.
1 code implementation • 9 Oct 2022 • Jiannan Xiang, Zhengzhong Liu, Yucheng Zhou, Eric P. Xing, Zhiting Hu
In the data disambiguation stage, we employ the prompted GPT-3 model to understand possibly ambiguous triples from the input data and convert each into a short sentence with reduced ambiguity.
no code implementations • 16 Jun 2022 • Yucheng Zhou, Tao Shen, Xiubo Geng, Chongyang Tao, Can Xu, Guodong Long, Binxing Jiao, Daxin Jiang
A ranker plays an indispensable role in the de facto 'retrieval & rerank' pipeline, but its training still lags behind -- learning from moderate negatives or/and serving as an auxiliary module for a retriever.
1 code implementation • ACL 2022 • Yucheng Zhou, Tao Shen, Xiubo Geng, Guodong Long, Daxin Jiang
Generating new events given context with correlated ones plays a crucial role in many event-centric reasoning tasks.
no code implementations • 13 Oct 2021 • Yucheng Zhou, Xiubo Geng, Tao Shen, Guodong Long, Daxin Jiang
Event correlation reasoning infers whether a natural language paragraph containing multiple events conforms to human common sense.
no code implementations • NAACL 2021 • Yucheng Zhou, Xiubo Geng, Tao Shen, Wenqiang Zhang, Daxin Jiang
That is, we can only access training data in a high-resource language, while need to answer multilingual questions without any labeled data in target languages.