no code implementations • ACL 2022 • Ling Liu, Mans Hulden
Annotation errors that stem from various sources are usually unavoidable when performing large-scale annotation of linguistic data.
no code implementations • EMNLP (insights) 2021 • Ling Liu, Mans Hulden
Backtranslation is a common technique for leveraging unlabeled data in low-resource scenarios in machine translation.
no code implementations • EMNLP 2020 • Sarah Moeller, Ling Liu, Changbing Yang, Katharina Kann, Mans Hulden
An intermediate step in the linguistic analysis of an under-documented language is to find and organize inflected forms that are attested in natural speech.
1 code implementation • 1 Mar 2025 • Tiansheng Huang, Sihao Hu, Fatih Ilhan, Selim Furkan Tekin, Zachary Yahn, Yichang Xu, Ling Liu
While safety alignment has been extensively studied for LLM, there is still a large research gap for Large Reasoning Models (LRMs) that equip with improved reasoning capability.
1 code implementation • 6 Feb 2025 • Selim Furkan Tekin, Fatih Ilhan, Tiansheng Huang, Sihao Hu, Zachary Yahn, Ling Liu
First, we develop an agent-fusion framework for encouraging multiple LLM based agents to collaborate in producing the final inference output for each LLM query.
1 code implementation • 29 Jan 2025 • Tiansheng Huang, Sihao Hu, Fatih Ilhan, Selim Furkan Tekin, Ling Liu
By designing a new red-teaming method, we in this paper show that purely relying on the moderation guardrail for data filtration is not reliable.
no code implementations • 12 Dec 2024 • Yi Luo, Linghang Shi, Yihao Li, Aobo Zhuang, Yeyun Gong, Ling Liu, Chen Lin
Conventional biomedical research is increasingly labor-intensive due to the exponential growth of scientific literature and datasets.
1 code implementation • 26 Nov 2024 • Selim Furkan Tekin, Fatih Ilhan, Tiansheng Huang, Sihao Hu, Zachary Yahn, Ling Liu
The former penalizes the selection errors of the expert-router, and the latter mediates the expert weights drifting during fine-tuning and dynamically adjusts the fusion behavior of the resulting model by canalizing the activations on the experts.
no code implementations • 11 Oct 2024 • Qin Liu, Chao Shang, Ling Liu, Nikolaos Pappas, Jie Ma, Neha Anna John, Srikanth Doss, Lluis Marquez, Miguel Ballesteros, Yassine Benajiba
The safety alignment ability of Vision-Language Models (VLMs) is prone to be degraded by the integration of the vision module compared to its LLM backbone.
1 code implementation • 4 Oct 2024 • Selim Furkan Tekin, Fatih Ilhan, Tiansheng Huang, Sihao Hu, Ling Liu
This paper presents LLM-TOPLA, a diversity-optimized LLM ensemble method with three unique properties: (i) We introduce the focal diversity metric to capture the diversity-performance correlation among component LLMs of an ensemble.
1 code implementation • 3 Oct 2024 • Aparna Elangovan, Lei Xu, Jongwoo Ko, Mahsa Elyasi, Ling Liu, Sravan Bodapati, Dan Roth
Specifically, we demonstrate that when the proportion of samples with variation or uncertainty in human assigned labels is relatively high, machine labels (generated by automatic evaluation methods) may superficially appear to have similar or better correlation with the human majority label compared to the human-to-human (HH) correlation.
2 code implementations • 26 Sep 2024 • Tiansheng Huang, Sihao Hu, Fatih Ilhan, Selim Furkan Tekin, Ling Liu
To clear up concern, this paper provide a comprehensive overview to three aspects of harmful fine-tuning: attacks setting, defense design and evaluation methodology.
1 code implementation • 3 Sep 2024 • Tiansheng Huang, Sihao Hu, Fatih Ilhan, Selim Furkan Tekin, Ling Liu
For the first time in the literature, we in this paper show that \textit{harmful perturbation} over the model weights should be the root cause of alignment-broken of harmful fine-tuning.
3 code implementations • 18 Aug 2024 • Tiansheng Huang, Gautam Bhattacharya, Pratik Joshi, Josh Kimball, Ling Liu
To this end, we propose Antidote, a post-fine-tuning stage solution, which remains \textbf{\textit{agnostic to the training hyper-parameters in the fine-tuning stage}}.
1 code implementation • 19 Jul 2024 • Ka-Ho Chow, Sihao Hu, Tiansheng Huang, Ling Liu
Second, we incorporate a perceptibility optimization to preserve the visual quality of the protected facial images.
1 code implementation • 28 May 2024 • Tiansheng Huang, Sihao Hu, Fatih Ilhan, Selim Furkan Tekin, Ling Liu
Recent studies show that Large Language Models (LLMs) with safety alignment can be jail-broken by fine-tuning on a dataset mixed with harmful data.
no code implementations • 28 May 2024 • Aparna Elangovan, Ling Liu, Lei Xu, Sravan Bodapati, Dan Roth
In this position paper, we argue that human evaluation of generative large language models (LLMs) should be a multidisciplinary undertaking that draws upon insights from disciplines such as user experience research and human behavioral psychology to ensure that the experimental design and results are reliable.
no code implementations • 5 Apr 2024 • Selim Furkan Tekin, Fatih Ilhan, Tiansheng Huang, Sihao Hu, Ka-Ho Chow, Margaret L. Loper, Ling Liu
This paper presents FusionShot, a focal diversity optimized few-shot ensemble learning approach for boosting the robustness and generalization performance of pre-trained few-shot models.
1 code implementation • 2 Apr 2024 • Sihao Hu, Tiansheng Huang, Gaowen Liu, Ramana Rao Kompella, Fatih Ilhan, Selim Furkan Tekin, Yichang Xu, Zachary Yahn, Ling Liu
The development of game agents holds a critical role in advancing towards Artificial General Intelligence.
1 code implementation • 2 Feb 2024 • Tiansheng Huang, Sihao Hu, Ling Liu
The new paradigm of finetuning-as-a-service introduces a new attack surface for Large Language Models (LLMs): a few harmful data uploaded by users can easily trick the finetuning to produce an alignment-broken model.
1 code implementation • 2 Feb 2024 • Sihao Hu, Tiansheng Huang, Ling Liu
We introduce PokeLLMon, the first LLM-embodied agent that achieves human-parity performance in tactical battle games, as demonstrated in Pokemon battles.
no code implementations • 2 Feb 2024 • Wenqi Wei, Ling Liu
Emerging Distributed AI systems are revolutionizing big data computing and data processing capabilities with growing economic and societal impact.
no code implementations • CVPR 2024 • Yun Liu, Haolin Yang, Xu Si, Ling Liu, Zipeng Li, Yuxiang Zhang, Yebin Liu, Li Yi
Humans commonly work with multiple objects in daily life and can intuitively transfer manipulation skills to novel objects by understanding object functional regularities.
1 code implementation • CVPR 2024 • Fatih Ilhan, Gong Su, Selim Furkan Tekin, Tiansheng Huang, Sihao Hu, Ling Liu
With the recent advances in vision transformers and large language models (LLMs) finetuning costly large models on downstream learning tasks poses significant challenges under limited computational resources.
1 code implementation • 17 Nov 2023 • Yanzhao Wu, Ka-Ho Chow, Wenqi Wei, Ling Liu
This paper presents a novel deep ensemble pruning approach, which can efficiently identify smaller deep ensembles and provide higher ensemble accuracy than the entire deep ensemble of a large number of member networks.
1 code implementation • 3 Oct 2023 • Yanzhao Wu, Ka-Ho Chow, Wenqi Wei, Ling Liu
We show that this two-tier heterogeneity driven ensemble construction method can compose an ensemble team that promotes high ensemble diversity and low negative correlation among member models of the ensemble, strengthening ensemble robustness against both negative examples and adversarial attacks.
1 code implementation • 2 Oct 2023 • Sihao Hu, Tiansheng Huang, Fatih İlhan, Selim Furkan Tekin, Ling Liu
The goal of auditor is to yield a broad spectrum of vulnerabilities with the hope of encompassing the correct answer, whereas the goal of critic that evaluates the validity of identified vulnerabilities is to minimize the number of false positives.
no code implementations • 20 Sep 2023 • Minhui Xue, Surya Nepal, Ling Liu, Subbu Sethuvenkatraman, Xingliang Yuan, Carsten Rudolph, Ruoxi Sun, Greg Eisenhauer
This paper plans to develop an Equitable and Responsible AI framework with enabling techniques and algorithms for the Internet of Energy (IoE), in short, RAI4IoE.
1 code implementation • 6 Sep 2023 • Sanjana Vijay Ganesh, Yanzhao Wu, Gaowen Liu, Ramana Kompella, Ling Liu
Object tracking is an important functionality of edge video analytic systems and services.
no code implementations • 19 Aug 2023 • Xigang Sun, Jingya Zhou, Ling Liu, Wenqi Wei
Predicting information cascade popularity is a fundamental problem in social networks.
1 code implementation • 29 May 2023 • Shengchao Liu, Jiongxiao Wang, Yijin Yang, Chengpeng Wang, Ling Liu, Hongyu Guo, Chaowei Xiao
This research sheds light on the potential of ChatGPT and conversational LLMs for drug editing.
1 code implementation • 25 May 2023 • Youssef Elmougy, Ling Liu
This enables both the detection of fraudulent transactions and the detection of illicit addresses (actors) in the Bitcoin network by leveraging four types of graph data: (i) the transaction-to-transaction graph, representing the money flow in the Bitcoin network, (ii) the address-to-address interaction graph, capturing the types of transaction flows between Bitcoin addresses, (iii) the address-transaction graph, representing the bi-directional money flow between addresses and transactions (BTC flow from input address to one or more transactions and BTC flow from a transaction to one or more output addresses), and (iv) the user entity graph, capturing clusters of Bitcoin addresses representing unique Bitcoin users.
no code implementations • 18 May 2023 • Sharon Levy, Neha Anna John, Ling Liu, Yogarshi Vyas, Jie Ma, Yoshinari Fujinuma, Miguel Ballesteros, Vittorio Castelli, Dan Roth
As a result, it is critical to examine biases within each language and attribute.
1 code implementation • 10 May 2023 • Wenqi Wei, Ling Liu, Jingya Zhou, Ka-Ho Chow, Yanzhao Wu
Next, we present a gradient leakage resilient approach to securing distributed SGD in federated learning, with differential privacy controlled noise as the tool.
1 code implementation • 29 Mar 2023 • Sihao Hu, Zhen Zhang, Bingqiao Luo, Shengliang Lu, Bingsheng He, Ling Liu
As various forms of fraud proliferate on Ethereum, it is imperative to safeguard against these malicious activities to protect susceptible users from being victimized.
1 code implementation • CVPR 2023 • Ka-Ho Chow, Ling Liu, Wenqi Wei, Fatih Ilhan, Yanzhao Wu
Based on the insights, we introduce a three-tier forensic framework to identify and expel Trojaned gradients and reclaim the performance over the course of FL.
1 code implementation • 15 Jan 2023 • Fatih Ilhan, Ka-Ho Chow, Sihao Hu, Tiansheng Huang, Selim Tekin, Wenqi Wei, Yanzhao Wu, Myungjin Lee, Ramana Kompella, Hugo Latapie, Gaowen Liu, Ling Liu
Instead of having every sample go through all DNN layers during prediction, EENet learns an early exit scheduler, which can intelligently terminate the inference earlier for certain predictions, which the model has high confidence of early exit.
1 code implementation • CVPR 2023 • Fatih Ilhan, Gong Su, Ling Liu
In most FL approaches, all edge clients are assumed to have sufficient computation capabilities to participate in the learning of a deep neural network (DNN) model.
1 code implementation • 21 Dec 2022 • Shengchao Liu, Weili Nie, Chengpeng Wang, Jiarui Lu, Zhuoran Qiao, Ling Liu, Jian Tang, Chaowei Xiao, Anima Anandkumar
Here we present a multi-modal molecule structure-text model, MoleculeSTM, by jointly learning molecules' chemical structures and textual descriptions via a contrastive learning strategy.
no code implementations • 30 Nov 2022 • Sunghyun Sim, Ling Liu, Hyerim Bae
Process mining is a methodology for the derivation and analysis of process models based on the event log.
no code implementations • 28 Nov 2022 • Imam Mustafa Kamal, Hyerim Bae, Ling Liu
With the emergence of deep learning, metric learning has gained significant popularity in numerous machine learning tasks dealing with complex and large-scale datasets, such as information retrieval, object recognition and recommendation systems.
1 code implementation • 24 Oct 2022 • Yanzhao Wu, Ling Liu
First, we develop an LR tuning mechanism for auto-verification of a given LR policy with respect to the desired accuracy goal under the pre-defined training time constraint.
no code implementations • 18 Oct 2022 • Shengjie Zheng, Ling Liu, Junjie Yang, Jianwei Zhang, Tao Su, Bin Yue, Xiaojian Li
The development of artificial intelligence (AI) and robotics are both based on the tenet of "science and technology are people-oriented", and both need to achieve efficient communication with the human brain.
no code implementations • 25 Dec 2021 • Wenqi Wei, Ling Liu
Although deep learning with differential privacy is a defacto standard for publishing deep learning models with differential privacy guarantee, we show that differentially private algorithms with fixed privacy parameters are vulnerable against gradient leakage attacks.
1 code implementation • IEEE International Conference on Data Mining (ICDM) 2021 • Yanzhao Wu, Ling Liu
Evaluated using two benchmark datasets, we show that the proposed focal diversity powered hierarchical pruning can find significantly smaller ensembles of deep neural network models while achieving the same or better classification generalizability.
1 code implementation • 22 Oct 2021 • Zhongwei Xie, Ling Liu, Yanzhao Wu, Luo Zhong, Lin Li
This paper introduces a two-phase deep feature engineering framework for efficient learning of semantics enhanced joint embedding, which clearly separates the deep feature engineering in data preprocessing from training the text-image joint embedding model.
no code implementations • 14 Oct 2021 • Jingya Zhou, Ling Liu, Wenqi Wei, Jianxi Fan
This survey paper reviews the design principles and the different node embedding techniques for network representation learning over homogeneous networks.
no code implementations • 9 Aug 2021 • Zhongwei Xie, Ling Liu, Lin Li, Luo Zhong
This paper presents a three-tier modality alignment approach to learning text-image joint embedding, coined as JEMA, for cross-modal retrieval of cooking recipes and food images.
1 code implementation • 2 Aug 2021 • Zhongwei Xie, Ling Liu, Yanzhao Wu, Lin Li, Luo Zhong
We present a Multi-modal Semantics enhanced Joint Embedding approach (MSJE) for learning a common feature space between the two modalities (text and image), with the ultimate goal of providing high-performance cross-modal retrieval services.
no code implementations • 2 Aug 2021 • Zhongwei Xie, Ling Liu, Lin Li, Luo Zhong
This paper introduces a two-phase deep feature calibration framework for efficient learning of semantics enhanced text-image cross-modal joint embedding, which clearly separates the deep feature calibration in data preprocessing from training the joint embedding model.
no code implementations • ACL 2021 • Sarah Moeller, Ling Liu, Mans Hulden
However, the importance and usefulness of POS tags needs to be examined as NLP expands to low-resource languages because linguists who provide many annotated resources do not place priority on early identification and tagging of POS.
2 code implementations • 27 Jul 2021 • Yanzhao Wu, Ling Liu, Ramana Kompella
A common performance requirement in these mission-critical edge services is the near real-time latency of online object detection on edge devices.
2 code implementations • 2 Jul 2021 • Wenqi Wei, Ling Liu, Yanzhao Wu, Gong Su, Arun Iyengar
This paper presents a gradient leakage resilient approach to privacy-preserving federated learning with per training example-based client differential privacy, coined as Fed-CDP.
1 code implementation • CVPR 2021 • Yanzhao Wu, Ling Liu, Zhongwei Xie, Ka-Ho Chow, Wenqi Wei
Our new metrics significantly improve the intrinsic correlation between high ensemble diversity and high ensemble accuracy.
no code implementations • 19 May 2021 • Ling Liu
Neural network approaches have been applied to computational morphology with great success, improving the performance of most tasks by a large margin and providing new perspectives for modeling.
no code implementations • 8 May 2021 • Jian Chen, Xuxin Zhang, Rui Zhang, Chen Wang, Ling Liu
The results demonstrate that De-Pois is effective and efficient for detecting poisoned data against all the four types of poisoning attacks, with both the accuracy and F1-score over 0. 9 on average.
no code implementations • ACL 2022 • Ling Liu, Mans Hulden
Deep learning sequence models have been successfully applied to the task of morphological inflection.
1 code implementation • 1 Jan 2021 • Yanzhao Wu, Ling Liu
(3) We introduce a two phase hierarchical pruning method to effectively identify and prune those deep ensembles with high HQ diversity scores, aiming to increase the lower and upper bounds on ensemble accuracy for the selected ensembles.
1 code implementation • COLING 2020 • Ling Liu, Mans Hulden
Analogy is assumed to be the cognitive mechanism speakers resort to in order to inflect an unknown form of a lexeme based on knowledge of other words in a language.
1 code implementation • 20 Oct 2020 • Yanzhao Wu, Ling Liu, Zhongwei Xie, Juhyun Bae, Ka-Ho Chow, Wenqi Wei
Ensemble learning is gaining renewed interests in recent years.
no code implementations • 14 Sep 2020 • Mehmet Emre Gursoy, Vivekanand Rajasekar, Ling Liu
Given a real trace dataset D, the differential privacy parameter epsilon controlling the strength of privacy protection, and the utility/error metric Err of interest; OptaTrace uses Bayesian optimization to optimize DPLTS such that the output error (measured in terms of given metric Err) is minimized while epsilon-differential privacy is satisfied.
no code implementations • 14 Sep 2020 • Wenqi Wei, Ling Liu
Third, XEnsemble provides a suite of algorithms to combine input verification and output verification to protect the DNN prediction models from both adversarial examples and out of distribution inputs.
2 code implementations • 16 Jul 2020 • Vale Tolpegin, Stacey Truex, Mehmet Emre Gursoy, Ling Liu
Federated learning (FL) is an emerging paradigm for distributed training of large-scale deep neural networks in which participants' data remains on their own devices with only model updates being shared with a central server.
no code implementations • 15 Jul 2020 • Wenqi Wei, Qi Zhang, Ling Liu
First, we explore three interesting properties between Bitcoin transaction accounts: topological connectivity pattern of Bitcoin accounts, transaction amount pattern, and transaction dynamics.
1 code implementation • 11 Jul 2020 • Ka-Ho Chow, Ling Liu, Mehmet Emre Gursoy, Stacey Truex, Wenqi Wei, Yanzhao Wu
We demonstrate that the proposed framework can serve as a methodical benchmark for analyzing adversarial behaviors and risks in real-time object detection systems.
no code implementations • WS 2020 • Ling Liu, Mans Hulden
This paper presents the submission by the CU Ling team from the University of Colorado to SIGMORPHON 2020 shared task 0 on morphological inflection.
no code implementations • 5 Jun 2020 • Stacey Truex, Ling Liu, Ka-Ho Chow, Mehmet Emre Gursoy, Wenqi Wei
However, in federated learning model parameter updates are collected iteratively from each participant and consist of high dimensional, continuous values with high precision (10s of digits after the decimal point), making existing LDP protocols inapplicable.
2 code implementations • 22 Apr 2020 • Wenqi Wei, Ling Liu, Margaret Loper, Ka-Ho Chow, Mehmet Emre Gursoy, Stacey Truex, Yanzhao Wu
FL offers default client privacy by allowing clients to keep their sensitive data on local devices and to only share local training parameter updates with the federated server.
2 code implementations • 9 Apr 2020 • Ka-Ho Chow, Ling Liu, Mehmet Emre Gursoy, Stacey Truex, Wenqi Wei, Yanzhao Wu
The rapid growth of real-time huge data capturing has pushed the deep learning and data analytic computing to the edge systems.
no code implementations • 3 Jan 2020 • Zhengping Liang, Weiqi Liang, Xiuju Xu, Ling Liu, Zexuan Zhu
Experimental results on multi-tasking multi-objective optimization test suites show that EMT-PD is superior to other six state-of-the-art evolutionary multi/single-tasking algorithms.
no code implementations • 21 Nov 2019 • Stacey Truex, Ling Liu, Mehmet Emre Gursoy, Wenqi Wei, Lei Yu
Second, through MPLens, we highlight how the vulnerability of pre-trained models under membership inference attack is not uniform across all classes, particularly when the training data itself is skewed.
no code implementations • 1 Oct 2019 • Wenqi Wei, Ling Liu, Margaret Loper, Ka-Ho Chow, Emre Gursoy, Stacey Truex, Yanzhao Wu
Deep neural network (DNN) has demonstrated its success in multiple domains.
no code implementations • 29 Aug 2019 • Ling Liu, Wenqi Wei, Ka-Ho Chow, Margaret Loper, Emre Gursoy, Stacey Truex, Yanzhao Wu
In this paper we first give an overview of the concept of ensemble diversity and examine the three types of ensemble diversity in the context of DNN classifiers.
no code implementations • 21 Aug 2019 • Ka-Ho Chow, Wenqi Wei, Yanzhao Wu, Ling Liu
Deep neural networks (DNNs) have demonstrated impressive performance on many challenging machine learning tasks.
1 code implementation • 18 Aug 2019 • Yanzhao Wu, Ling Liu, Juhyun Bae, Ka-Ho Chow, Arun Iyengar, Calton Pu, Wenqi Wei, Lei Yu, Qi Zhang
Learning Rate (LR) is an important hyper-parameter to tune for effective training of deep neural networks (DNNs).
no code implementations • 15 May 2019 • Mehmet Emre Gursoy, Acar Tamersoy, Stacey Truex, Wenqi Wei, Ling Liu
In this paper, we address the small user population problem by introducing the concept of Condensed Local Differential Privacy (CLDP) as a specialization of LDP, and develop a suite of CLDP protocols that offer desirable statistical utility while preserving privacy.
Cryptography and Security Databases
no code implementations • 3 Apr 2019 • Lei Yu, Ling Liu, Calton Pu, Mehmet Emre Gursoy, Stacey Truex
However, when the training datasets are crowdsourced from individuals and contain sensitive information, the model parameters may encode private information and bear the risks of privacy leakage.
1 code implementation • 29 Oct 2018 • Yanzhao Wu, Ling Liu, Calton Pu, Wenqi Cao, Semih Sahin, Wenqi Wei, Qi Zhang
Big data powered Deep Learning (DL) and its applications have blossomed in recent years, fueled by three technological trends: a large amount of digitized data openly accessible, a growing number of DL software frameworks in open source and commercial markets, and a selection of affordable parallel computing hardware devices.
no code implementations • COLING 2018 • Miikka Silfverberg, Ling Liu, Mans Hulden
In supervised learning of morphological patterns, the strategy of generalizing inflectional tables into more abstract paradigms through alignment of the longest common subsequence found in an inflection table has been proposed as an efficient method to deduce the inflectional behavior of unseen word forms.
no code implementations • 29 Jun 2018 • Wenqi Wei, Ling Liu, Margaret Loper, Stacey Truex, Lei Yu, Mehmet Emre Gursoy, Yanzhao Wu
The burgeoning success of deep learning has raised the security and privacy concerns as more and more tasks are accompanied with sensitive data.
1 code implementation • 28 Jun 2018 • Stacey Truex, Ling Liu, Mehmet Emre Gursoy, Lei Yu, Wenqi Wei
Our empirical results additionally show that (1) using the type of target model under attack within the attack model may not increase attack effectiveness and (2) collaborative learning in federated systems exposes vulnerabilities to membership inference risks when the adversary is a participant in the federation.
Cryptography and Security