1 code implementation • 1 May 2025 • Bang Zhang, Ruotian Ma, Qingxuan Jiang, Peisong Wang, Jiaqi Chen, Zheng Xie, Xingyu Chen, Yue Wang, Fanghua Ye, Jian Li, Yifan Yang, Zhaopeng Tu, Xiaolong Li
Assessing how well a large language model (LLM) understands human, rather than merely text, remains an open challenge.
no code implementations • 20 Mar 2025 • Ruihan Yang, Fanghua Ye, Jian Li, Siyu Yuan, Yikai Zhang, Zhaopeng Tu, Xiaolong Li, Deqing Yang
In this work, we introduce Critique-Guided Improvement (CGI), a novel two-player framework, comprising an actor model that explores an environment and a critic model that generates detailed nature language feedback.
no code implementations • 20 Feb 2025 • Jing Xiong, Jianghan Shen, Chuanyang Zheng, Zhongwei Wan, Chenyang Zhao, Chiwun Yang, Fanghua Ye, Hongxia Yang, Lingpeng Kong, Ngai Wong
To mitigate the attention sink issue, we propose an attention calibration strategy that reduces biases, ensuring more stable long-range attention.
no code implementations • 28 Nov 2024 • Shenghui Li, Edith C. -H. Ngai, Fanghua Ye, Thiemo Voigt
This paper introduces a novel security threat to FedPEFT, termed PEFT-as-an-Attack (PaaA), which exposes how PEFT can be exploited as an attack vector to circumvent PLMs' safety alignment and generate harmful content in response to malicious prompts.
no code implementations • 4 Oct 2024 • Jing Xiong, Jianghan Shen, Fanghua Ye, Chaofan Tao, Zhongwei Wan, Jianqiao Lu, Xun Wu, Chuanyang Zheng, Zhijiang Guo, Lingpeng Kong, Ngai Wong
By grouping layers and heads based on their uncertainty, UNComp adaptively compresses both the hidden states and the KV cache.
no code implementations • 3 Oct 2024 • Zixuan Li, Jing Xiong, Fanghua Ye, Chuanyang Zheng, Xun Wu, Jianqiao Lu, Zhongwei Wan, Xiaodan Liang, Chengming Li, Zhenan Sun, Lingpeng Kong, Ngai Wong
We present UncertaintyRAG, a novel approach for long-context Retrieval-Augmented Generation (RAG) that utilizes Signal-to-Noise Ratio (SNR)-based span uncertainty to estimate similarity between text chunks.
1 code implementation • 24 Jul 2024 • Anhao Zhao, Fanghua Ye, Jinlan Fu, Xiaoyu Shen
Recent research presents two conflicting views on ICL: One emphasizes the impact of similar examples in the demonstrations, stressing the need for label correctness and more shots.
no code implementations • 18 Jun 2024 • Shenghui Li, Fanghua Ye, Meng Fang, Jiaxu Zhao, Yun-Hin Chan, Edith C. -H. Ngai, Thiemo Voigt
The recent development of Foundation Models (FMs), represented by large language models, vision transformers, and multimodal models, has been making a significant impact on both academia and industry.
no code implementations • 16 May 2024 • Jiahuan Pei, Irene Viola, Haochen Huang, Junxiao Wang, Moonisa Ahsan, Fanghua Ye, Jiang Yiming, Yao Sai, Di Wang, Zhumin Chen, Pengjie Ren, Pablo Cesar
We present a demonstration of a multimodal fine-grained training assistant for LEGO brick assembly in a pilot XR environment.
1 code implementation • 12 Feb 2024 • Jianhui Pang, Fanghua Ye, Derek Fai Wong, Xin He, Wanshun Chen, Longyue Wang
Large language models (LLMs) predominantly employ decoder-only transformer architectures, necessitating the retention of keys/values information for historical tokens to provide contextual information and avoid redundant computation.
1 code implementation • 23 Jan 2024 • Fanghua Ye, Mingming Yang, Jianhui Pang, Longyue Wang, Derek F. Wong, Emine Yilmaz, Shuming Shi, Zhaopeng Tu
The proliferation of open-source Large Language Models (LLMs) from various institutions has highlighted the urgent need for comprehensive evaluation methods.
1 code implementation • 16 Jan 2024 • Jianhui Pang, Fanghua Ye, Longyue Wang, Dian Yu, Derek F. Wong, Shuming Shi, Zhaopeng Tu
This study revisits these challenges, offering insights into their ongoing relevance in the context of advanced Large Language Models (LLMs): domain mismatch, amount of parallel data, rare word prediction, translation of long sentences, attention model as word alignment, and sub-optimal beam search.
no code implementations • 14 Dec 2023 • Shitong Sun, Fanghua Ye, Shaogang Gong
Composed image retrieval attempts to retrieve an image of interest from gallery images through a composed query of a reference image and its corresponding modified text.
1 code implementation • 23 Oct 2023 • Zihan Zhang, Meng Fang, Fanghua Ye, Ling Chen, Mohammad-Reza Namazi-Rad
Dialogue state tracking (DST) plays an important role in task-oriented dialogue systems.
1 code implementation • 15 Oct 2023 • Fanghua Ye, Meng Fang, Shenghui Li, Emine Yilmaz
Furthermore, we propose distilling the rewriting capabilities of LLMs into smaller models to reduce rewriting latency.
1 code implementation • 21 May 2023 • Fanghua Ye, Zhiyuan Hu, Emine Yilmaz
It assumes that the performance of a dialogue system can be measured by user satisfaction and uses an estimator to simulate users.
1 code implementation • 22 Oct 2022 • Fanghua Ye, Xi Wang, Jie Huang, Shenghui Li, Samuel Stern, Emine Yilmaz
Experimental results demonstrate that all three schemes can achieve competitive performance.
no code implementations • ACL 2022 • Yue Feng, Aldo Lipani, Fanghua Ye, Qiang Zhang, Emine Yilmaz
Existing approaches that have considered such relations generally fall short in: (1) fusing prior slot-domain membership relations and dialogue-aware dynamic slot relations explicitly, and (2) generalizing to unseen domains.
1 code implementation • Findings (ACL) 2022 • Fanghua Ye, Yue Feng, Emine Yilmaz
In this paper, instead of improving the annotation quality further, we propose a general framework, named ASSIST (lAbel noiSe-robuSt dIalogue State Tracking), to train DST models robustly from noisy labels.
1 code implementation • SIGDIAL (ACL) 2022 • Fanghua Ye, Jarana Manotumruksa, Emine Yilmaz
The annotations in the training set remain unchanged (same as MultiWOZ 2. 1) to elicit robust and noise-resilient model training.
1 code implementation • 22 Jan 2021 • Fanghua Ye, Jarana Manotumruksa, Qiang Zhang, Shenghui Li, Emine Yilmaz
Then a stacked slot self-attention is applied on these features to learn the correlations among slots.
1 code implementation • 14 Jan 2021 • Shenghui Li, Edith Ngai, Fanghua Ye, Thiemo Voigt
In this paper, we address this challenge by proposing Auto-weighted Robust Federated Learning (arfl), a novel approach that jointly learns the global model and the weights of local updates to provide robustness against corrupted data sources.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Fanghua Ye, Jarana Manotumruksa, Emine Yilmaz
Semantic hashing is a powerful paradigm for representing texts as compact binary hash codes.
1 code implementation • 1 Jun 2020 • Fanghua Ye, Zhiwei Lin, Chuan Chen, Zibin Zheng, Hong Huang
The proliferation of Web services makes it difficult for users to select the most appropriate one among numerous functionally identical or similar service candidates.
2 code implementations • CIKM 2018 • Fanghua Ye, Chuan Chen, Zibin Zheng
Considering the complicated and diversified topology structures of real-world networks, it is highly possible that the mapping between the original network and the community membership space contains rather complex hierarchical information, which cannot be interpreted by classic shallow NMF-based approaches.
Ranked #1 on
Node Classification
on Wiki