1 code implementation • EMNLP 2020 • Qiongkai Xu, Lizhen Qu, Zeyu Gao, Gholamreza Haffari
In this work, we propose to protect personal information by warning users of detected suspicious sentences generated by conversational assistants.
1 code implementation • 17 Feb 2025 • Yuncheng Hua, Lizhen Qu, Zhuang Li, Hao Xue, Flora D. Salim, Gholamreza Haffari
This paper proposes a low-cost, tuning-free method using in-context learning (ICL) to enhance LLM alignment.
no code implementations • 8 Feb 2025 • Manh Luong, Khai Nguyen, Dinh Phung, Gholamreza Haffari, Lizhen Qu
However, the contrastive method ignores the temporal information when measuring similarity across acoustic and linguistic modalities, leading to inferior performance.
1 code implementation • 31 Oct 2024 • Hao Yang, Lizhen Qu, Ehsan Shareghi, Gholamreza Haffari
Our results under these settings demonstrate that open-source audio LMMs suffer an average attack success rate of 69. 14% on harmful audio questions, and exhibit safety vulnerabilities when distracted with non-speech audio noise.
no code implementations • 16 Oct 2024 • Minghao Wu, Thuy-Trang Vu, Lizhen Qu, Gholamreza Haffari
In this paper, we introduce GraphFilter, a novel method that represents the dataset as a bipartite graph, linking sentences to their constituent n-grams.
1 code implementation • 15 Oct 2024 • Hao Yang, Lizhen Qu, Ehsan Shareghi, Gholamreza Haffari
Moreover, JSP achieves a state-of-the-art attack success rate of 92% on GPT-4 on the harmful query benchmark, and exhibits strong resistant to defence strategies.
no code implementations • 4 Oct 2024 • Shilin Qu, Weiqing Wang, Xin Zhou, Haolan Zhan, Zhuang Li, Lizhen Qu, Linhao Luo, Yuan-Fang Li, Gholamreza Haffari
Our empirical results show: (i) the quality of the SCNs derived from synthetic data is comparable to that from real dialogues annotated with gold frames, and (ii) the quality of the SCNs extracted from real data, annotated with either silver (predicted) or gold frames, surpasses that without the frame annotations.
no code implementations • 16 Sep 2024 • Zhixi Cai, Cristian Rojas Cardenas, Kevin Leo, Chenyuan Zhang, Kal Backman, Hanbing Li, Boying Li, Mahsa Ghorbanali, Stavya Datta, Lizhen Qu, Julian Gutierrez Santiago, Alexey Ignatiev, Yuan-Fang Li, Mor Vered, Peter J Stuckey, Maria Garcia de la Banda, Hamid Rezatofighi
This paper addresses the problem of autonomous UAV search missions, where a UAV must locate specific Entities of Interest (EOIs) within a time limit, based on brief descriptions in large, hazard-prone environments with keep-out zones.
no code implementations • 5 Sep 2024 • Xianbing Zhao, Lizhen Qu, Tao Feng, Jianfei Cai, Buzhou Tang
This work proposes a novel and simple sequential learning strategy to train models on videos and texts for multimodal sentiment analysis.
no code implementations • 29 Jul 2024 • Tao Feng, Lizhen Qu, Niket Tandon, Zhuang Li, Xiaoxi Kang, Gholamreza Haffari
Recent advances in artificial intelligence have seen Large Language Models (LLMs) demonstrate notable proficiency in causal discovery tasks.
1 code implementation • 25 Jun 2024 • Hao Yang, Lizhen Qu, Ehsan Shareghi, Gholamreza Haffari
Based on the taxonomy, we create a small-scale dataset for evaluating current LMMs capability in detecting these categories of risk.
no code implementations • 25 Jun 2024 • Tao Feng, Lizhen Qu, Xiaoxi Kang, Gholamreza Haffari
Automatically evaluating the quality of responses in open-domain dialogue systems is a challenging but crucial task.
no code implementations • 19 Jun 2024 • Xiaoxi Kang, Lizhen Qu, Lay-Ki Soon, Zhuang Li, Adnan Trakic
This paper introduces LEGALSEMI, a benchmark specifically curated for legal scenario analysis.
no code implementations • 18 Jun 2024 • Yuncheng Hua, Yujin Huang, Shuo Huang, Tao Feng, Lizhen Qu, Chris Bain, Richard Bassed, Gholamreza Haffari
Inspired by causal discovery, we propose a novel deep latent model in the variational autoencoder (VAE) framework, which not only captures the underlying latent structures of data but also utilizes the easily transferable knowledge of emotions as the bridge to link the distributions of events in different domains.
1 code implementation • 16 Jun 2024 • Zhuang Li, Yuncheng Hua, Thuy-Trang Vu, Haolan Zhan, Lizhen Qu, Gholamreza Haffari
Recent studies emphasize that manually ensuring a consistent response style and maintaining high data quality in training sets can significantly improve the performance of fine-tuned Large Language Models (LLMs) while reducing the number of training examples needed.
no code implementations • 13 Jun 2024 • Minghao Wu, Thuy-Trang Vu, Lizhen Qu, Gholamreza Haffari
In this work, we propose a general, model-agnostic, reinforcement learning framework, Mixture-of-Skills (MoS), that learns to optimize data usage automatically during the fine-tuning process.
no code implementations • 6 Jun 2024 • Shuo Huang, William MacLean, Xiaoxi Kang, Anqi Wu, Lizhen Qu, Qiongkai Xu, Zhuang Li, Xingliang Yuan, Gholamreza Haffari
Increasing concerns about privacy leakage issues in academia and industry arise when employing NLP models from third-party providers to process sensitive texts.
1 code implementation • 16 May 2024 • Manh Luong, Khai Nguyen, Nhat Ho, Reza Haf, Dinh Phung, Lizhen Qu
The Learning-to-match (LTM) framework proves to be an effective inverse optimal transport approach for learning the underlying ground metric between two sources of data, facilitating subsequent matching.
1 code implementation • 10 May 2024 • Ilia Kuznetsov, Osama Mohammed Afzal, Koen Dercksen, Nils Dycke, Alexander Goldberg, Tom Hope, Dirk Hovy, Jonathan K. Kummerfeld, Anne Lauscher, Kevin Leyton-Brown, Sheng Lu, Mausam, Margot Mieskes, Aurélie Névéol, Danish Pruthi, Lizhen Qu, Roy Schwartz, Noah A. Smith, Thamar Solorio, Jingyan Wang, Xiaodan Zhu, Anna Rogers, Nihar B. Shah, Iryna Gurevych
We hope that our work will help set the agenda for research in machine-assisted scientific quality control in the age of AI, within the NLP community and beyond.
1 code implementation • 21 Apr 2024 • Tao Feng, Lizhen Qu, Zhuang Li, Haolan Zhan, Yuncheng Hua, Gholamreza Haffari
Machine learning models have made incredible progress, but they still struggle when applied to examples from unseen domains.
1 code implementation • CVPR 2024 • Chuang Lin, Yi Jiang, Lizhen Qu, Zehuan Yuan, Jianfei Cai
To address it, we formulate object detection as a generative problem and propose a simple framework named GenerateU, which can detect dense objects and generate their names in a free-form way.
no code implementations • 10 Mar 2024 • Zhuo Zhang, Jingyuan Zhang, Jintao Huang, Lizhen Qu, Hongzhi Zhang, Qifan Wang, Xun Zhou, Zenglin Xu
Federated instruction tuning (FedIT) has emerged as a promising solution, by consolidating collaborative training across multiple data owners, thereby resulting in a privacy-preserving learning model.
no code implementations • 17 Feb 2024 • Haolan Zhan, Zhuang Li, Xiaoxi Kang, Tao Feng, Yuncheng Hua, Lizhen Qu, Yi Ying, Mei Rianto Chandra, Kelly Rosalin, Jureynolds Jureynolds, Suraj Sharma, Shilin Qu, Linhao Luo, Lay-Ki Soon, Zhaleh Semnani Azad, Ingrid Zukerman, Gholamreza Haffari
While collecting sufficient human-authored data is costly, synthetic conversations provide suitable amounts of data to help mitigate the scarcity of training data, as well as the chance to assess the alignment between LLMs and humans in the awareness of social norms.
no code implementations • 2 Feb 2024 • Haolan Zhan, YuFei Wang, Tao Feng, Yuncheng Hua, Suraj Sharma, Zhuang Li, Lizhen Qu, Zhaleh Semnani Azad, Ingrid Zukerman, Gholamreza Haffari
Negotiation is a crucial ability in human communication.
no code implementations • 29 Jan 2024 • Yuncheng Hua, Lizhen Qu, Gholamreza Haffari
We introduce a simple tuning-free and label-free In-Context Learning (ICL) method to identify high-quality ICL exemplars for the remediator, where we propose a novel select criteria, called value impact, to measure the quality of the negotiation outcomes.
1 code implementation • 29 Jan 2024 • Yuncheng Hua, Zhuang Li, Linhao Luo, Kadek Ananta Satriadi, Tao Feng, Haolan Zhan, Lizhen Qu, Suraj Sharma, Ingrid Zukerman, Zhaleh Semnani-Azad, Gholamreza Haffari
We have released our code and software at:~\url{https://github. com/AnonymousEACLDemo/SADAS}.
no code implementations • 27 Jan 2024 • Minghao Wu, YuFei Wang, George Foster, Lizhen Qu, Gholamreza Haffari
Document-level neural machine translation (DocNMT) aims to generate translations that are both coherent and cohesive, in contrast to its sentence-level counterpart.
no code implementations • 12 Jan 2024 • Minghao Wu, Thuy-Trang Vu, Lizhen Qu, George Foster, Gholamreza Haffari
We provide an in-depth analysis of these LLMs tailored for DocMT, examining translation errors, discourse phenomena, strategies for training and inference, the data efficiency of parallel documents, recent test set evaluations, and zero-shot crosslingual transfer.
1 code implementation • 8 Dec 2023 • Tongxin Hu, Zhuang Li, Xin Jin, Lizhen Qu, Xin Zhang
Annually, e-commerce platforms incur substantial financial losses due to trademark infringements, making it crucial to identify and mitigate potential legal risks tied to merchant information registered to the platforms.
1 code implementation • 23 Oct 2023 • Xiaoxi Kang, Lizhen Qu, Lay-Ki Soon, Adnan Trakic, Terry Yue Zhuo, Patrick Charles Emerton, Genevieve Grant
Each scenario in the corpus is annotated with a complete IRAC analysis in a semi-structured format so that both machines and legal professionals are able to interpret and understand the annotations.
1 code implementation • 27 May 2023 • Zhuang Li, Yuyang Chai, Terry Yue Zhuo, Lizhen Qu, Gholamreza Haffari, Fei Li, Donghong Ji, Quan Hung Tran
Textual scene graph parsing has become increasingly important in various vision-language applications, including image caption evaluation and image retrieval.
Ranked #2 on
Human Judgment Correlation
on Flickr8k-Expert
no code implementations • 22 May 2023 • Zhuang Li, Lizhen Qu, Philip R. Cohen, Raj V. Tumuluri, Gholamreza Haffari
Multilingual semantic parsing aims to leverage the knowledge from the high-resource languages to improve low-resource semantic parsing, yet commonly suffers from the data imbalance problem.
1 code implementation • 8 May 2023 • Bhanu Prakash Voutharoja, Lizhen Qu, Fatemeh Shiri
Our model parses a form into a word-relation graph in order to identify entities and relations jointly and reduce the time complexity of inference.
1 code implementation • 2 May 2023 • Haolan Zhan, Sameen Maruf, Lizhen Qu, YuFei Wang, Ingrid Zukerman, Gholamreza Haffari
Flowchart-grounded troubleshooting dialogue (FTD) systems, which follow the instructions of a flowchart to diagnose users' problems in specific domains (e. g., vehicle, laptop), have been gaining research interest in recent years.
1 code implementation • 24 Apr 2023 • Haolan Zhan, Zhuang Li, YuFei Wang, Linhao Luo, Tao Feng, Xiaoxi Kang, Yuncheng Hua, Lizhen Qu, Lay-Ki Soon, Suraj Sharma, Ingrid Zukerman, Zhaleh Semnani-Azad, Gholamreza Haffari
To the best of our knowledge, SocialDial is the first socially-aware dialogue dataset that covers multiple social factors and has fine-grained labels.
Cultural Vocal Bursts Intensity Prediction
Synthetic Data Generation
1 code implementation • 2 Mar 2023 • Tao Feng, Lizhen Qu, Gholamreza Haffari
In this paper, we conduct the first study on spurious correlations for open-domain response generation models based on a corpus CGDIALOG curated in our work.
no code implementations • 16 Feb 2023 • Minghao Wu, George Foster, Lizhen Qu, Gholamreza Haffari
Existing work in document-level neural machine translation commonly concatenates several consecutive sentences as a pseudo-document, and then learns inter-sentential dependencies.
1 code implementation • 20 Dec 2022 • Zhuo Zhang, Yuanhang Yang, Yong Dai, Lizhen Qu, Zenglin Xu
To facilitate the research of PETuning in FL, we also develop a federated tuning framework FedPETuning, which allows practitioners to exploit different PETuning methods under the FL training paradigm conveniently.
no code implementations • 18 Dec 2022 • Haolan Zhan, YuFei Wang, Tao Feng, Yuncheng Hua, Suraj Sharma, Zhuang Li, Lizhen Qu, Gholamreza Haffari
Negotiation is one of the crucial abilities in human communication, and there has been a resurgent research interest in negotiation dialogue systems recently, which goal is to empower intelligent agents with such ability that can efficiently help humans resolve conflicts or reach beneficial agreements.
1 code implementation • 27 Nov 2022 • Chuang Lin, Peize Sun, Yi Jiang, Ping Luo, Lizhen Qu, Gholamreza Haffari, Zehuan Yuan, Jianfei Cai
In this paper, we propose a novel open-vocabulary object detection framework directly learning from image-text pair data.
no code implementations • 11 Oct 2022 • Terry Yue Zhuo, Yaqing Liao, Yuecheng Lei, Lizhen Qu, Gerard de Melo, Xiaojun Chang, Yazhou Ren, Zenglin Xu
We introduce ViLPAct, a novel vision-language benchmark for human activity planning.
1 code implementation • 27 Feb 2022 • Zhuang Li, Lizhen Qu, Qiongkai Xu, Tongtong Wu, Tianyang Zhan, Gholamreza Haffari
In this paper, we propose a variational autoencoder with disentanglement priors, VAE-DPRIOR, for task-specific natural language generation with none or a handful of task-specific labeled examples.
1 code implementation • 10 Nov 2021 • Chuang Lin, Yi Jiang, Jianfei Cai, Lizhen Qu, Gholamreza Haffari, Zehuan Yuan
Vision-and-Language Navigation (VLN) is a task that an agent is required to follow a language instruction to navigate to the goal position, which relies on the ongoing interactions with the environment during moving.
no code implementations • Findings (EMNLP) 2021 • Sheng Bi, Xiya Cheng, Yuan-Fang Li, Lizhen Qu, Shirong Shen, Guilin Qi, Lu Pan, Yinlin Jiang
The ability to generate natural-language questions with controlled complexity levels is highly desirable as it further expands the applicability of question generation.
1 code implementation • EMNLP 2021 • Zhuang Li, Lizhen Qu, Gholamreza Haffari
We conduct extensive experiments to study the research problems involved in continual semantic parsing and demonstrate that a neural semantic parser trained with TotalRecall achieves superior performance than the one trained directly with the SOTA continual learning algorithms and achieve a 3-6 times speedup compared to re-training from scratch.
no code implementations • COLING 2022 • Qiongkai Xu, Xuanli He, Lingjuan Lyu, Lizhen Qu, Gholamreza Haffari
Machine-learning-as-a-service (MLaaS) has attracted millions of users to their splendid large-scale models.
no code implementations • ACL 2021 • Farhad Moghimifar, Lizhen Qu, Yue Zhuo, Gholamreza Haffari, Mahsa Baktashmotlagh
The dynamic nature of commonsense knowledge postulates models capable of performing multi-hop reasoning over new situations.
no code implementations • EACL 2021 • Shuo Huang, Zhuang Li, Lizhen Qu, Lei Pan
In this paper, we provide the empirical study on the robustness of semantic parsers in the presence of adversarial attacks.
1 code implementation • EACL 2021 • Zhuang Li, Lizhen Qu, Shuo Huang, Gholamreza Haffari
In this work, we investigate the problems of semantic parsing in a few-shot learning setting.
1 code implementation • COLING 2020 • Farhad Moghimifar, Lizhen Qu, Yue Zhuo, Mahsa Baktashmotlagh, Gholamreza Haffari
However, current approaches in this realm lack the ability to perform commonsense reasoning upon facing an unseen situation, mostly due to incapability of identifying a diverse range of implicit social relations.
1 code implementation • COLING 2020 • Zhuang Li, Lizhen Qu, Gholamreza Haffari
Semantic parsing is the task of translating natural language utterances into machine-readable meaning representations.
no code implementations • WS 2019 • Qiongkai Xu, Lizhen Qu, Chenchen Xu, Ran Cui
Biased decisions made by automatic systems have led to growing concerns in research communities.
1 code implementation • IJCNLP 2019 • Qiongkai Xu, Chenchen Xu, Lizhen Qu
In this paper, we describe ALTER, an auxiliary text rewriting tool that facilitates the rewriting process for natural language generation tasks, such as paraphrasing, text simplification, fairness-aware text rewriting, and text style transfer.
no code implementations • ICLR 2019 • Tue Le, Tuan Nguyen, Trung Le, Dinh Phung, Paul Montague, Olivier De Vel, Lizhen Qu
Due to the sharp increase in the severity of the threat imposed by software vulnerabilities, the detection of vulnerabilities in binary code has become an important concern in the software industry, such as the embedded systems industry, and in the field of computer security.
no code implementations • 13 Aug 2018 • Qiongkai Xu, Juyan Zhang, Lizhen Qu, Lexing Xie, Richard Nock
In this paper, we investigate the diversity aspect of paraphrase generation.
1 code implementation • NeurIPS 2017 • Richard Nock, Zac Cranko, Aditya Krishna Menon, Lizhen Qu, Robert C. Williamson
In this paper, we unveil a broad class of distributions for which such convergence happens --- namely, deformed exponential families, a wide superset of exponential families --- and show tight connections with the three other key GAN parameters: loss, game and architecture.
no code implementations • ACL 2017 • Sunghwan Mac Kim, Qiongkai Xu, Lizhen Qu, Stephen Wan, C{\'e}cile Paris
In social media, demographic inference is a critical task in order to gain a better understanding of a cohort and to facilitate interacting with one{'}s audience.
no code implementations • 24 Jan 2017 • Qiongkai Xu, Qing Wang, Chenchen Xu, Lizhen Qu
In this paper, we propose a graph-based recursive neural network framework for collective vertex classification.
no code implementations • 20 Dec 2016 • Shijie Zhang, Lizhen Qu, ShaoDi You, Zhenglu Yang, Jiawan Zhang
In this paper, we propose the first model to be able to generate visually grounded questions with diverse types for a single image.
no code implementations • EMNLP 2016 • Lizhen Qu, Gabriela Ferraro, Liyuan Zhou, Weiwei Hou, Timothy Baldwin
In named entity recognition, we often don't have a large in-domain training corpus or a knowledge base with adequate coverage to train a model directly.
2 code implementations • CVPR 2017 • Giorgio Patrini, Alessandro Rozza, Aditya Menon, Richard Nock, Lizhen Qu
We present a theoretically grounded approach to train deep neural networks, including recurrent networks, subject to class-dependent label noise.
Ranked #2 on
Image Classification
on Clothing1M (using clean data)
(using extra training data)
1 code implementation • NAACL 2016 • Dat Quoc Nguyen, Kairit Sirts, Lizhen Qu, Mark Johnson
Knowledge bases of real-world facts about entities and their relationships are useful resources for a variety of natural language processing tasks.
no code implementations • CONLL 2016 • Dat Quoc Nguyen, Kairit Sirts, Lizhen Qu, Mark Johnson
Knowledge bases are useful resources for many natural language processing tasks, however, they are far from complete.
no code implementations • 21 Apr 2015 • Lizhen Qu, Gabriela Ferraro, Liyuan Zhou, Weiwei Hou, Nathan Schneider, Timothy Baldwin
Word embeddings -- distributed word representations that can be learned from unlabelled data -- have been shown to have high utility in many natural language processing applications.
no code implementations • 4 Aug 2014 • Lizhen Qu, Bjoern Andres
Estimating (learning) a maximally probable measure, given (a training set of) related and unrelated pairs, is a convex optimization problem.
no code implementations • TACL 2014 • Lizhen Qu, Yi Zhang, Rui Wang, Lili Jiang, Rainer Gemulla, Gerhard Weikum
Extracting instances of sentiment-oriented relations from user-generated web documents is important for online marketing analysis.