no code implementations • 6 Mar 2025 • Yijie Xu, Aiwei Liu, Xuming Hu, Lijie Wen, Hui Xiong
Our experiments reveal that backdoor watermarking could effectively detect IP Violation, while inference-time watermark distillation is applicable in both scenarios but less robust to further fine-tuning and has a more significant impact on LLM performance compared to backdoor watermarking.
1 code implementation • 17 Feb 2025 • Leyi Pan, Aiwei Liu, Shiyu Huang, Yijian Lu, Xuming Hu, Lijie Wen, Irwin King, Philip S. Yu
The radioactive nature of Large Language Model (LLM) watermarking enables the detection of watermarks inherited by student models when trained on the outputs of watermarked teacher models, making it a promising tool for preventing unauthorized knowledge distillation.
no code implementations • 10 Dec 2024 • Bo Li, Shaolin Zhu, Lijie Wen
Image Translation (IT) holds immense potential across diverse domains, enabling the translation of textual content within images into various languages.
no code implementations • 22 Nov 2024 • Junzhe Chen, Tianshu Zhang, Shiyu Huang, Yuwei Niu, Linfeng Zhang, Lijie Wen, Xuming Hu
Despite the recent breakthroughs achieved by Large Vision Language Models (LVLMs) in understanding and responding to complex visual-textual contexts, their inherent hallucination tendencies limit their practical application in real-world scenarios that demand high levels of precision.
no code implementations • 19 Nov 2024 • Xiaohe Li, Feilong Huang, Zide Fan, Fangli Mou, Leilei Lin, Yingyan Hou, Lijie Wen
Trajectory prediction for multi-agents in complex scenarios is crucial for applications like autonomous driving.
no code implementations • 6 Oct 2024 • Aiwei Liu, Haoping Bai, Zhiyun Lu, Yanchao Sun, Xiang Kong, Simon Wang, Jiulong Shan, Albin Madappally Jose, Xiaojiang Liu, Lijie Wen, Philip S. Yu, Meng Cao
In this work, we propose that the optimal data for DPO has equal expected rewards for each token in winning and losing responses, as there is no difference in token importance.
1 code implementation • 4 Oct 2024 • Aiwei Liu, Sheng Guan, Yiming Liu, Leyi Pan, Yifei Zhang, Liancheng Fang, Lijie Wen, Philip S. Yu, Xuming Hu
Finally, we propose that the key to enhancing the imperceptibility of watermarked LLMs is to increase the randomness of watermark key selection.
1 code implementation • 2 Oct 2024 • Zitian Gao, Boye Niu, Xuzheng He, Haotian Xu, Hongzhang Liu, Aiwei Liu, Xuming Hu, Lijie Wen
Thus, we conducted extensive ablation studies and quantitative analysis on components of MCTS, revealing the impact of each component on the MCTS reasoning performance of LLMs.
1 code implementation • 8 Sep 2024 • Leyi Pan, Aiwei Liu, Yijian Lu, Zitian Gao, Yichen Di, Shiyu Huang, Lijie Wen, Irwin King, Philip S. Yu
Watermarking algorithms for large language models (LLMs) have attained high accuracy in detecting LLM-generated text.
1 code implementation • 11 Jun 2024 • Shiao Meng, Xuming Hu, Aiwei Liu, Fukun Ma, Yawen Yang, Shuang Li, Lijie Wen
To this end, we systematically investigate the robustness of DocRE models to entity name variations in this work.
1 code implementation • 1 Jun 2024 • Xuan Wu, Di Wang, Lijie Wen, Yubin Xiao, Chunguo Wu, Yuesong Wu, Chaoyu Yu, Douglas L. Maskell, You Zhou
Although several surveys on Neural Combinatorial Optimization (NCO) solvers specifically designed to solve Vehicle Routing Problems (VRPs) have been conducted.
1 code implementation • 16 May 2024 • Leyi Pan, Aiwei Liu, Zhiwei He, Zitian Gao, Xuandong Zhao, Yijian Lu, Binglin Zhou, Shuliang Liu, Xuming Hu, Lijie Wen, Irwin King, Philip S. Yu
However, the abundance of LLM watermarking algorithms, their intricate mechanisms, and the complex evaluation procedures and perspectives pose challenges for researchers and the community to easily experiment with, understand, and assess the latest advancements.
no code implementations • 29 Mar 2024 • Shuang Li, Jiahua Wang, Lijie Wen
Multi-modal reasoning plays a vital role in bridging the gap between textual and visual information, enabling a deeper understanding of the context.
no code implementations • 15 Mar 2024 • Yiyang Luo, Ke Lin, Chao Gu, Jiahui Hou, Lijie Wen, Ping Luo
The proliferation of large language models (LLMs) in generating content raises concerns about text copyright.
no code implementations • 5 Mar 2024 • Yutong Li, Lu Chen, Aiwei Liu, Kai Yu, Lijie Wen
In this work, we firstly focus on the independent literature summarization step and introduce ChatCite, an LLM agent with human workflow guidance for comparative literature summary.
no code implementations • 1 Mar 2024 • Bo Li, Qinghua Zhao, Lijie Wen
Previous works have evaluated memorization by comparing model outputs with training corpora, examining how factors such as data duplication, model size, and prompt length influence memorization.
1 code implementation • 26 Feb 2024 • Junzhe Chen, Xuming Hu, Shuodi Liu, Shiyu Huang, Wei-Wei Tu, Zhaofeng He, Lijie Wen
Recent advancements in large language models (LLMs) have revealed their potential for achieving autonomous agents possessing human-level intelligence.
no code implementations • 25 Feb 2024 • Xuming Hu, Xiaochuan Li, Junzhe Chen, Yinghui Li, Yangning Li, Xiaoguang Li, Yasheng Wang, Qun Liu, Lijie Wen, Philip S. Yu, Zhijiang Guo
To this end, we propose evaluating the robustness of generative search engines in the realistic and high-risk setting, where adversaries have only black-box system access and seek to deceive the model into returning incorrect responses.
1 code implementation • 19 Feb 2024 • Aiwei Liu, Haoping Bai, Zhiyun Lu, Xiang Kong, Simon Wang, Jiulong Shan, Meng Cao, Lijie Wen
In this paper, we propose a method to evaluate the response preference by using the output probabilities of response pairs under contrastive prompt pairs, which could achieve better performance on LLaMA2-7B and LLaMA2-13B compared to RLAIF.
no code implementations • 13 Feb 2024 • Xiaohe Li, Feilong Huang, Zide Fan, Fangli Mou, Yingyan Hou, Chen Qian, Lijie Wen
Trajectory prediction has garnered widespread attention in different fields, such as autonomous driving and robotic navigation.
no code implementations • 13 Dec 2023 • Aiwei Liu, Leyi Pan, Yijian Lu, Jingjing Li, Xuming Hu, Xi Zhang, Lijie Wen, Irwin King, Hui Xiong, Philip S. Yu
This paper conducts a comprehensive survey of the current state of text watermarking technology, covering four main aspects: (1) an overview and comparison of different text watermarking techniques; (2) evaluation methods for text watermarking algorithms, including their detectability, impact on text or LLM quality, robustness under target or untargeted attacks; (3) potential application scenarios for text watermarking technology; (4) current challenges and future directions for text watermarking.
1 code implementation • 25 Oct 2023 • Xuming Hu, Junzhe Chen, Aiwei Liu, Shiao Meng, Lijie Wen, Philip S. Yu
Additionally, our method is orthogonal to previous multimodal fusions, and using it on prior SOTA fusions further improves 5. 47% F1.
1 code implementation • 24 Oct 2023 • Shiao Meng, Xuming Hu, Aiwei Liu, Shu'ang Li, Fukun Ma, Yawen Yang, Lijie Wen
However, existing works often struggle to obtain class prototypes with accurate relational semantics: 1) To build prototype for a target relation type, they aggregate the representations of all entity pairs holding that relation, while these entity pairs may also hold other relations, thus disturbing the prototype.
2 code implementations • 10 Oct 2023 • Aiwei Liu, Leyi Pan, Xuming Hu, Shiao Meng, Lijie Wen
In this work, we propose a semantic invariant watermarking method for LLMs that provides both attack robustness and security robustness.
no code implementations • 8 Oct 2023 • Xuming Hu, Junzhe Chen, Xiaochuan Li, Yufei Guo, Lijie Wen, Philip S. Yu, Zhijiang Guo
Large language models (LLMs) have recently driven striking performance improvements across a range of natural language processing tasks.
3 code implementations • 30 Jul 2023 • Aiwei Liu, Leyi Pan, Xuming Hu, Shu'ang Li, Lijie Wen, Irwin King, Philip S. Yu
Experiments demonstrate that our algorithm attains high detection accuracy and computational efficiency through neural networks.
no code implementations • 29 May 2023 • Aiwei Liu, Wei Liu, Xuming Hu, Shuang Li, Fukun Ma, Yawen Yang, Lijie Wen
Based on these observations, we propose a method named \texttt{p-align} to improve the compositional generalization of Text-to-SQL models.
1 code implementation • 22 May 2023 • Shuang Li, Xuming Hu, Aiwei Liu, Yawen Yang, Fukun Ma, Philip S. Yu, Lijie Wen
In this paper, we propose a novel Soft prompt learning framework with the Multilingual Verbalizer (SoftMV) for XNLI.
Cross-Lingual Natural Language Inference
Cross-Lingual Transfer
no code implementations • 12 May 2023 • Yawen Yang, Xuming Hu, Fukun Ma, Shu'ang Li, Aiwei Liu, Lijie Wen, Philip S. Yu
Existing works for nested NER ignore the recognition order and boundary position relation of nested entities.
1 code implementation • 2 May 2023 • Xuming Hu, Zhaochen Hong, Zhijiang Guo, Lijie Wen, Philip S. Yu
In light of this, we propose a fact verification model named ReRead to retrieve evidence and verify claim that: (1) Train the evidence retriever to obtain interpretable evidence (i. e., faithfulness and plausibility criteria); (2) Train the claim verifier to revisit the evidence retrieved by the optimized evidence retriever to improve the accuracy.
1 code implementation • 12 Mar 2023 • Aiwei Liu, Xuming Hu, Lijie Wen, Philip S. Yu
This paper presents the first comprehensive analysis of ChatGPT's Text-to-SQL ability.
no code implementations • 11 Nov 2022 • Xuming Hu, Shiao Meng, Chenwei Zhang, Xiangli Yang, Lijie Wen, Irwin King, Philip S. Yu
Low-Resource Information Extraction (LRIE) strives to use unsupervised data, reducing the required resources and human annotation.
1 code implementation • 31 Oct 2022 • Aiwei Liu, Honghai Yu, Xuming Hu, Shu'ang Li, Li Lin, Fukun Ma, Yawen Yang, Lijie Wen
We propose the first character-level white-box adversarial attack method against transformer models.
no code implementations • 19 Oct 2022 • Xuming Hu, Yong Jiang, Aiwei Liu, Zhongqiang Huang, Pengjun Xie, Fei Huang, Lijie Wen, Philip S. Yu
Data augmentation techniques have been used to alleviate the problem of scarce labeled data in various NER tasks (flat, nested, and discontinuous NER tasks).
no code implementations • COLING 2022 • Xuming Hu, Zhijiang Guo, Yu Fu, Lijie Wen, Philip S. Yu
A scene graph is a semantic representation that expresses the objects, attributes, and relationships between objects in a scene.
1 code implementation • 8 Aug 2022 • Aiwei Liu, Xuming Hu, Li Lin, Lijie Wen
First, we extract a schema linking graph from PLMs through a probing procedure in an unsupervised manner.
1 code implementation • NAACL 2022 • Xuming Hu, Zhijiang Guo, Guanyu Wu, Aiwei Liu, Lijie Wen, Philip S. Yu
The explosion of misinformation spreading in the media ecosystem urges for automated fact-checking.
no code implementations • 31 May 2022 • Shu'ang Li, Xuming Hu, Li Lin, Aiwei Liu, Lijie Wen, Philip S. Yu
Natural Language Inference (NLI) is a growingly essential task in natural language understanding, which requires inferring the relationship between the sentence pairs (premise and hypothesis).
1 code implementation • NAACL 2022 • Xuming Hu, Shuliang Liu, Chenwei Zhang, Shu`ang Li, Lijie Wen, Philip S. Yu
Unsupervised relation extraction aims to extract the relationship between entities from natural language sentences without prior information on relational scope or distribution.
no code implementations • 5 Feb 2022 • Xiaohe Li, Lijie Wen, Yawen Deng, Fuli Feng, Xuming Hu, Lei Wang, Zide Fan
Graph Neural Network (GNN) is an emerging technique for graph-based learning tasks such as node classification.
no code implementations • 26 Jan 2022 • Shu'ang Li, Xuming Hu, Li Lin, Lijie Wen
We adopt a cross attention module to learn the joint representations of the sentence pairs.
no code implementations • 18 Jan 2022 • Li Lin, Yixin Cao, Lifu Huang, Shu'ang Li, Xuming Hu, Lijie Wen, Jianmin Wang
To alleviate the knowledge forgetting issue, we design two modules, Im and Gm, for each type of knowledge, which are combined via prompt tuning.
1 code implementation • EMNLP 2021 • Xuming Hu, Chenwei Zhang, Yawen Yang, Xiaohe Li, Li Lin, Lijie Wen, Philip S. Yu
Low-resource Relation Extraction (LRE) aims to extract relation facts from limited labeled corpora when human annotation is scarce.
1 code implementation • ACL 2021 • Chen Qian, Fuli Feng, Lijie Wen, Chunping Ma, Pengjun Xie
In inference, given a factual input document, Corsair imagines its two counterfactual counterparts to distill and mitigate the two biases captured by the poisonous model.
2 code implementations • Findings (ACL) 2021 • Chenyao Liu, Shengnan An, Zeqi Lin, Qian Liu, Bei Chen, Jian-Guang Lou, Lijie Wen, Nanning Zheng, Dongmei Zhang
In this paper, we propose LeAR, an end-to-end neural model to learn algebraic recombination for compositional generalization.
Ranked #2 on
Semantic Parsing
on CFQ
no code implementations • 23 Dec 2020 • Xiaohe Li, Lijie Wen, Chen Qian, Jianmin Wang
Heterogeneous network embedding aims to embed nodes into low-dimensional vectors which capture rich intrinsic information of heterogeneous networks.
no code implementations • COLING 2020 • Xingyao Zhang, Linjun Shou, Jian Pei, Ming Gong, Lijie Wen, Daxin Jiang
The abundant semi-structured data on the Web, such as HTML-based tables and lists, provide commercial search engines a rich information source for question answering (QA).
1 code implementation • Findings (EMNLP) 2021 • Xuming Hu, Chenwei Zhang, Fukun Ma, Chenyao Liu, Lijie Wen, Philip S. Yu
To alleviate human efforts from obtaining large-scale annotations, Semi-Supervised Relation Extraction methods aim to leverage unlabeled data in addition to learning from limited samples.
1 code implementation • EMNLP 2020 • Xuming Hu, Chenwei Zhang, Yusong Xu, Lijie Wen, Philip S. Yu
Open relation extraction is the task of extracting open-domain relation facts from natural language sentences.
no code implementations • 16 May 2019 • Chen Qian, Lijie Wen, Akhil Kumar
Process consistency checking (PCC), an interdiscipline of natural language processing (NLP) and business process management (BPM), aims to quantify the degree of (in)consistencies between graphical and textual descriptions of a process.
1 code implementation • 16 May 2019 • Chen Qian, Lijie Wen, Akhil Kumar, Leilei Lin, Li Lin, Zan Zong, Shuang Li, Jian-Min Wang
Process model extraction (PME) is a recently emerged interdiscipline between natural language processing (NLP) and business process management (BPM), which aims to extract process models from textual descriptions.