1 code implementation • ACL 2022 • Rongzhi Zhang, Yue Yu, Pranav Shetty, Le Song, Chao Zhang
Weakly-supervised learning (WSL) has shown promising results in addressing label scarcity on many NLP tasks, but manually designing a comprehensive, high-quality labeling rule set is tedious and difficult.
1 code implementation • NAACL 2022 • Yue Yu, Lingkai Kong, Jieyu Zhang, Rongzhi Zhang, Chao Zhang
We develop AcTune, a new framework that improves the label efficiency of active PLM fine-tuning by unleashing the power of unlabeled data via self-training.
no code implementations • 25 Nov 2024 • Yue Yu, Zhengxing Chen, Aston Zhang, Liang Tan, Chenguang Zhu, Richard Yuanzhe Pang, Yundi Qian, Xuewei Wang, Suchin Gururangan, Chao Zhang, Melanie Kambadur, Dhruv Mahajan, Rui Hou
Reward modeling is crucial for aligning large language models (LLMs) with human preferences, especially in reinforcement learning from human feedback (RLHF).
1 code implementation • 23 Nov 2024 • Tanghaoran Zhang, Yue Yu, XinJun Mao, Shangwen Wang, Kang Yang, Yao Lu, Zhang Zhang, Yuxin Zhao
Our result indicates that their adaptation ability is weaker than generation, with a nearly 15% decrease on pass@1 and more context-related errors.
no code implementations • 14 Nov 2024 • Zichen Liu, Yue Yu, Hao Ouyang, Qiuyu Wang, Ka Leong Cheng, Wen Wang, Zhiheng Liu, Qifeng Chen, Yujun Shen
Image editing involves a variety of complex tasks and requires efficient and precise manipulation techniques.
no code implementations • 28 Oct 2024 • Yue Yu, Prayag Tiwari
The framework simulates a collaborative team setting by training each model to specialize in distinct roles: Macro Analysts, Micro analysts, and Quantitative Analysts.
no code implementations • 27 Oct 2024 • Yiming Fan, Habib Najm, Yue Yu, Stewart Silling, Marta D'Elia
To develop a robust and reliable bottom-up homogenization framework, we propose a new framework, which we coin Embedded Nonlocal Operator Regression (ENOR), to learn a nonlocal homogenized surrogate model and its structural model error.
no code implementations • 3 Oct 2024 • Ning Liu, Lu Zhang, Tian Gao, Yue Yu
Specifically, we introduce DisentangO, a novel hyper-neural operator architecture designed to unveil and disentangle the latent physical factors of variation embedded within the black-box neural operator parameters.
no code implementations • 26 Sep 2024 • Zhengan Huang, Gongxian Zeng, Xin Mu, Yu Wang, Yue Yu
In this paper, we initiate the study of \emph{multi-designated detector watermarking (MDDW)} for large language models (LLMs).
no code implementations • 30 Aug 2024 • Ting Bai, Le Huang, Yue Yu, Cheng Yang, Cheng Hou, Zhe Zhao, Chuan Shi
A novel two-stage prompt-tuning MTL framework (MPT-Rec) is proposed to address task irrelevance and training efficiency problems in multi-task recommender systems.
no code implementations • 16 Aug 2024 • Zhongjian Zhang, Xiao Wang, Huichi Zhou, Yue Yu, Mengmei Zhang, Cheng Yang, Chuan Shi
By presenting the empirical results, we find that despite that LLMs can improve the robustness of GNNs, there is still an average decrease of 23. 1% in accuracy, implying that the GNNs remain extremely vulnerable against topology attack.
no code implementations • 14 Aug 2024 • Yue Yu, Ning Liu, Fei Lu, Tian Gao, Siavash Jafarzadeh, Stewart Silling
In this work, we propose a novel neural operator architecture based on the attention mechanism, which we coin Nonlocal Attention Operator (NAO), and explore its capability towards developing a foundation physical model.
no code implementations • 3 Jul 2024 • Ning Liu, Siavash Jafarzadeh, Brian Y. Lattimer, Shuna Ni, Jim Lua, Yue Yu
Our framework features a two-phase training strategy: (1) utilizing the large-in-amount while less accurate synthetic data for supervised pretraining, and (2) finetuning the phase-1 model with limited experimental data.
no code implementations • 2 Jul 2024 • Yue Yu, Wei Ping, Zihan Liu, Boxin Wang, Jiaxuan You, Chao Zhang, Mohammad Shoeybi, Bryan Catanzaro
Large language models (LLMs) typically utilize the top-k contexts from a retriever in retrieval-augmented generation (RAG).
Ranked #3 on Question Answering on PubMedQA
1 code implementation • 23 Jun 2024 • Haorui Wang, Marta Skreta, Cher-Tian Ser, Wenhao Gao, Lingkai Kong, Felix Strieth-Kalthoff, Chenru Duan, Yuchen Zhuang, Yue Yu, Yanqiao Zhu, Yuanqi Du, Alán Aspuru-Guzik, Kirill Neklyudov, Chao Zhang
Molecular discovery, when formulated as an optimization problem, presents significant computational challenges because optimization objectives can be non-differentiable.
no code implementations • 18 Jun 2024 • Jiayi Lin, Yutao Xie, Yue Yu, Yibiao Yang, Lei Zhang
While these models acquire vast amounts of code knowledge, they perform poorly on code understanding tasks, such as code search and clone detection, as they are specifically trained for generation.
1 code implementation • 16 Jun 2024 • Dongping Chen, Yue Huang, Siyuan Wu, Jingyu Tang, Liuyi Chen, Yilin Bai, Zhigang He, Chenlong Wang, Huichi Zhou, Yiqiang Li, Tianshuo Zhou, Yue Yu, Chujie Gao, Qihui Zhang, Yi Gui, Zhen Li, Yao Wan, Pan Zhou, Jianfeng Gao, Lichao Sun
We evaluate the capabilities of current state-of-the-art MLLMs, including ImageLLMs and VideoLLMs, in understanding various types of GUI content, especially dynamic and sequential content.
1 code implementation • 9 Jun 2024 • Xiaoqi Qiu, Yongjie Wang, Xu Guo, Zhiwei Zeng, Yue Yu, Yuhong Feng, Chunyan Miao
Counterfactually Augmented Data (CAD) involves creating new data samples by applying minimal yet sufficient modifications to flip the label of existing data samples to other classes.
1 code implementation • 5 Jun 2024 • Yuchen Zhuang, Haotian Sun, Yue Yu, Rushi Qiang, Qifan Wang, Chao Zhang, Bo Dai
To address these challenges, we propose HYDRA, a model factorization framework that captures both user-specific behavior patterns from historical data and shared general knowledge among all users to deliver personalized generation.
no code implementations • 23 May 2024 • Yuanzhao Zhai, Zhuo Zhang, Kele Xu, Hanyang Peng, Yue Yu, Dawei Feng, Cheng Yang, Bo Ding, Huaimin Wang
To overcome this limitation, we propose Online Self-Preferring (OSP) language models to learn from self-generated response pairs and self-judged preference strengths.
1 code implementation • 20 May 2024 • Ermo Hua, Biqing Qi, Kaiyan Zhang, Yue Yu, Ning Ding, Xingtai Lv, Kai Tian, BoWen Zhou
To obtain a unified understanding, we interpret SFT and PO with two sub-processes -- Preference Estimation and Transition Optimization -- defined at token level within the Markov Decision Process (MDP) framework.
no code implementations • 13 May 2024 • Ning Liu, Xuxiao Li, Manoj R. Rajanna, Edward W. Reutzel, Brady Sawyer, Prahalada Rao, Jim Lua, Nam Phan, Yue Yu
In terms of Laser Powder Bed Fusion (L-PBF) based additive manufacturing (AM), a DT can predict the current and future states of the melt pool and the resulting defects corresponding to the input laser parameters, evolve itself by assimilating in-situ sensor data, and optimize the laser parameters to mitigate defect formation.
1 code implementation • 5 May 2024 • Wenqi Shi, ran Xu, Yuchen Zhuang, Yue Yu, Haotian Sun, Hang Wu, Carl Yang, May D. Wang
Faced with the challenges of balancing model performance, computational resources, and data privacy, MedAdapter provides an efficient, privacy-preserving, cost-effective, and transparent solution for adapting LLMs to the biomedical domain.
1 code implementation • 29 Apr 2024 • ran Xu, Wenqi Shi, Yue Yu, Yuchen Zhuang, Yanqiao Zhu, May D. Wang, Joyce C. Ho, Chao Zhang, Carl Yang
Developing effective biomedical retrieval models is important for excelling at knowledge-intensive biomedical tasks but still challenging due to the deficiency of sufficient publicly annotated biomedical data and computational resources.
no code implementations • 17 Apr 2024 • Zichen Liu, Yihao Meng, Hao Ouyang, Yue Yu, Bolin Zhao, Daniel Cohen-Or, Huamin Qu
Through quantitative and qualitative evaluations, we demonstrate the effectiveness of our framework in generating coherent text animations that faithfully interpret user prompts while maintaining readability.
no code implementations • 27 Mar 2024 • Siavash Jafarzadeh, Stewart Silling, Lu Zhang, Colton Ross, Chung-Hao Lee, S. M. Rakibur Rahman, Shuodao Wang, Yue Yu
Our goal is to learn a nonlocal constitutive law together with the material microstructure, in the form of a heterogeneous fiber orientation field, from load-displacement field measurements.
1 code implementation • 18 Mar 2024 • Xiaojie Li, Yibo Yang, Xiangtai Li, Jianlong Wu, Yue Yu, Bernard Ghanem, Min Zhang
To tackle these challenges, we present GenView, a controllable framework that augments the diversity of positive views leveraging the power of pretrained generative models while preserving semantics.
1 code implementation • 17 Mar 2024 • Yuzhao Heng, Chunyuan Deng, Yitong Li, Yue Yu, Yinghao Li, Rongzhi Zhang, Chao Zhang
Although Large Language Models (LLMs) exhibit remarkable adaptability across domains, these models often fall short in structured knowledge extraction tasks such as named entity recognition (NER).
1 code implementation • 14 Mar 2024 • Mingya Zhang, Yue Yu, Limei Gu, Tingsheng Lin, Xianping Tao
In the field of medical image segmentation, models based on both CNN and Transformer have been thoroughly investigated.
no code implementations • 11 Mar 2024 • Shaojie Dai, Xin Liu, Ping Luo, Yue Yu
Large language model (LLM) has achieved promising performance in multilingual machine translation tasks through zero/few-shot prompts or prompt-tuning.
1 code implementation • 25 Feb 2024 • ran Xu, Wenqi Shi, Yue Yu, Yuchen Zhuang, Bowen Jin, May D. Wang, Joyce C. Ho, Carl Yang
We present RAM-EHR, a Retrieval AugMentation pipeline to improve clinical predictions on Electronic Health Records (EHRs).
1 code implementation • 22 Feb 2024 • Bin Liang, Ang Li, Jingqian Zhao, Lin Gui, Min Yang, Yue Yu, Kam-Fai Wong, Ruifeng Xu
Stance detection is a challenging task that aims to identify public opinion from social media platforms with respect to specific targets.
no code implementations • 22 Feb 2024 • Han Zhang, Lin Gui, Yu Lei, Yuanzhao Zhai, Yehong Zhang, Yulan He, Hui Wang, Yue Yu, Kam-Fai Wong, Bin Liang, Ruifeng Xu
Reinforcement Learning from Human Feedback (RLHF) is commonly utilized to improve the alignment of Large Language Models (LLMs) with human preferences.
no code implementations • 21 Feb 2024 • Lingxi Zhang, Yue Yu, Kuan Wang, Chao Zhang
Retrieval-augmented generation enhances large language models (LLMs) by incorporating relevant information from external knowledge sources.
1 code implementation • 19 Feb 2024 • Zhongjian Zhang, Mengmei Zhang, Yue Yu, Cheng Yang, Jiawei Liu, Chuan Shi
Furthermore, with GraphPAR, we quantify whether the fairness of each node is provable, i. e., predictions are always fair within a certain range of sensitive attribute semantics.
no code implementations • 24 Jan 2024 • Vidit Jain, Mukund Rungta, Yuchen Zhuang, Yue Yu, Zeyu Wang, Mu Gao, Jeffrey Skolnick, Chao Zhang
The best-performing models aim to learn a static representation by combining document and hierarchical label information.
1 code implementation • 13 Jan 2024 • Wenqi Shi, ran Xu, Yuchen Zhuang, Yue Yu, Jieyu Zhang, Hang Wu, Yuanda Zhu, Joyce Ho, Carl Yang, May D. Wang
Large language models (LLMs) have demonstrated exceptional capabilities in planning and tool utilization as autonomous agents, but few have been developed for medical problem-solving.
no code implementations • 11 Jan 2024 • Siavash Jafarzadeh, Stewart Silling, Ning Liu, Zhongqiang Zhang, Yue Yu
In this work, we introduce a novel integral neural operator architecture called the Peridynamic Neural Operator (PNO) that learns a nonlocal constitutive law from data.
no code implementations • 5 Jan 2024 • Jiahang Zhou, Yanyu Chen, Zicong Hong, Wuhui Chen, Yue Yu, Tao Zhang, Hui Wang, Chuanfu Zhang, Zibin Zheng
Additionally, the paper summarizes the challenges and presents a perspective on the future development direction of foundation model systems.
no code implementations • 1 Jan 2024 • Zhenyu Gao, Yue Yu, Qinshuang Wei, Ufuk Topcu, John-Paul Clarke
Urban air mobility (UAM), a transformative concept for the transport of passengers and cargo, faces several integration challenges in complex urban environments.
1 code implementation • 1 Jan 2024 • Jinglong Luo, Yehong Zhang, Zhuo Zhang, JiaQi Zhang, Xin Mu, Hui Wang, Yue Yu, Zenglin Xu
However, the application of SMPC in Privacy-Preserving Inference (PPI) for large language models, particularly those based on the Transformer architecture, often leads to considerable slowdowns or declines in performance.
no code implementations • 30 Dec 2023 • Yuanzhao Zhai, Han Zhang, Yu Lei, Yue Yu, Kele Xu, Dawei Feng, Bo Ding, Huaimin Wang
Reinforcement learning from human feedback (RLHF) emerges as a promising paradigm for aligning large language models (LLMs).
1 code implementation • 20 Dec 2023 • Naiyu Yin, Tian Gao, Yue Yu, Qiang Ji
We then propose an effective two-phase iterative DAG learning algorithm to address the increasing optimization difficulties and to learn a causal DAG from data with heteroscedastic variable noise under varying variance.
1 code implementation • 20 Dec 2023 • Chenglu Pan, Jiarong Xu, Yue Yu, Ziqi Yang, Qingbiao Wu, Chunping Wang, Lei Chen, Yang Yang
Extensive experiments show that our model achieves the best trade-off between accuracy and the fairness of model gradient, as well as superior payoff fairness.
no code implementations • 19 Dec 2023 • Xin Mu, Yu Wang, Zhengan Huang, Junzuo Lai, Yehong Zhang, Hui Wang, Yue Yu
In the rapidly growing digital economy, protecting intellectual property (IP) associated with digital products has become increasingly important.
1 code implementation • 18 Dec 2023 • Ning Liu, Yiming Fan, Xianyi Zeng, Milan Klöwer, Lu Zhang, Yue Yu
In this work, we introduce conservation law-encoded neural operators (clawNOs), a suite of NOs that endow inference with automatic satisfaction of such conservation laws.
1 code implementation • 4 Dec 2023 • Zhengyu Hu, Jieyu Zhang, Yue Yu, Yuchen Zhuang, Hui Xiong
This paper presents LEMR (Label-Efficient Model Ranking) and introduces the MoraBench Benchmark.
no code implementations • 14 Nov 2023 • Jing Nathan Yan, Tianqi Liu, Justin T Chiu, Jiaming Shen, Zhen Qin, Yue Yu, Yao Zhao, Charu Lakshmanan, Yair Kurzion, Alexander M. Rush, Jialu Liu, Michael Bendersky
Comparative reasoning plays a crucial role in text preference prediction; however, large language models (LLMs) often demonstrate inconsistencies in their reasoning.
no code implementations • 13 Nov 2023 • Yue Yu, Jiaming Shen, Tianqi Liu, Zhen Qin, Jing Nathan Yan, Jialu Liu, Chao Zhang, Michael Bendersky
To fully unleash the power of explanations, we propose EASE, an Explanation-Aware Soft Ensemble framework to empower in-context learning with LLMs.
1 code implementation • 1 Nov 2023 • ran Xu, Hejie Cui, Yue Yu, Xuan Kan, Wenqi Shi, Yuchen Zhuang, Wei Jin, Joyce Ho, Carl Yang
Clinical natural language processing requires methods that can address domain-specific challenges, such as complex medical terminology and clinical contexts.
no code implementations • 16 Oct 2023 • Yingwei Ma, Yue Yu, Shanshan Li, Yu Jiang, Yong Guo, Yuanliang Zhang, Yutao Xie, Xiangke Liao
Meanwhile, while traditional techniques leveraging such semantic information require complex static or dynamic code analysis to obtain features such as data flow and control flow, SeCoT demonstrates that this process can be fully automated via the intrinsic capabilities of LLMs (i. e., in-context learning), while being generalizable and applicable to challenging domains.
1 code implementation • 28 Sep 2023 • Yingwei Ma, Yue Liu, Yue Yu, Yuanliang Zhang, Yu Jiang, Changjian Wang, Shanshan Li
Inspired by the great success of code data in training LLMs, we naturally wonder at which training stage introducing code data can really help LLMs reasoning.
1 code implementation • NeurIPS 2023 • Yue Yu, Xiao Wang, Mengmei Zhang, Nian Liu, Chuan Shi
To this end, we propose the PrOvable Training (POT) for GCL, which regularizes the training of GCL to encode node embeddings that follows the GCL principle better.
no code implementations • 15 Aug 2023 • William Ward, Yue Yu, Jacob Levy, Negar Mehr, David Fridovich-Keil, Ufuk Topcu
We formulate an inverse learning problem in a Stackelberg game between a leader and a follower, where each player's action is the trajectory of a dynamical system.
no code implementations • 7 Aug 2023 • Zicong Hong, Xiaoyu Qiu, Jian Lin, Wuhui Chen, Yue Yu, Hui Wang, Song Guo, Wen Gao
Therefore, in this article, we present the concept of an intelligence-endogenous management platform for CNCs called \emph{CNC brain} based on artificial intelligence technologies.
no code implementations • 4 Aug 2023 • Xin Mu, Yu Wang, Yehong Zhang, JiaQi Zhang, Hui Wang, Yang Xiang, Yue Yu
Understanding the life cycle of the machine learning (ML) model is an intriguing area of research (e. g., understanding where the model comes from, how it is trained, and how it is used).
no code implementations • 5 Jul 2023 • Shiyu Liu, Shaogao Lv, Dun Zeng, Zenglin Xu, Hui Wang, Yue Yu
Federated learning is a decentralized and privacy-preserving technique that enables multiple clients to collaborate with a server to learn a global model without exposing their private data.
1 code implementation • NeurIPS 2023 • Yue Yu, Yuchen Zhuang, Jieyu Zhang, Yu Meng, Alexander Ratner, Ranjay Krishna, Jiaming Shen, Chao Zhang
Large language models (LLMs) have been recently leveraged as training data generators for various natural language processing (NLP) tasks.
no code implementations • 26 Jun 2023 • Jinglong Luo, Yehong Zhang, JiaQi Zhang, Shuang Qin, Hui Wang, Yue Yu, Zenglin Xu
In contrast to existing studies that protect the data privacy of GPR via homomorphic encryption, differential privacy, or federated learning, our proposed method is more practical and can be used to preserve the data privacy of both the model inputs and outputs for various data-sharing scenarios (e. g., horizontally/vertically-partitioned data).
1 code implementation • NeurIPS 2023 • Yuchen Zhuang, Yue Yu, Kuan Wang, Haotian Sun, Chao Zhang
To address this issue, we introduce a new dataset called ToolQA, which is designed to faithfully evaluate LLMs' ability to use external tools for question answering.
2 code implementations • 14 Jun 2023 • Yinghao Li, Lingkai Kong, Yuanqi Du, Yue Yu, Yuchen Zhuang, Wenhao Mu, Chao Zhang
While some studies have included UQ to improve molecular pre-trained models, the process of selecting suitable backbone and UQ methods for reliable molecular uncertainty estimation remains underexplored.
1 code implementation • 12 Jun 2023 • ran Xu, Yue Yu, Joyce C. Ho, Carl Yang
To address this challenge, we propose a weakly-supervised approach for scientific document classification using label names only.
no code implementations • 7 Jun 2023 • Carl Yang, Hejie Cui, Jiaying Lu, Shiyu Wang, ran Xu, Wenjing Ma, Yue Yu, Shaojun Yu, Xuan Kan, Chen Ling, Tianfan Fu, Liang Zhao, Joyce Ho, Fei Wang
This work aims to serve as a valuable resource for understanding the potential and opportunities of HKG in health research.
no code implementations • 5 Jun 2023 • Xuan Kan, Zimu Li, Hejie Cui, Yue Yu, ran Xu, Shaojun Yu, Zilong Zhang, Ying Guo, Carl Yang
Biological networks are commonly used in biomedical and healthcare domains to effectively model the structure of complex biological systems with interactions linking biological entities.
no code implementations • 5 Jun 2023 • Rongzhi Zhang, Yue Yu, Jiaming Shen, Xiquan Cui, Chao Zhang
In this work, we show that the standard implementation of the convex combination of base learners can hardly work due to the presence of noisy labels.
1 code implementation • 30 May 2023 • Yuchen Zhuang, Yue Yu, Lingkai Kong, Xiang Chen, Chao Zhang
Most existing methods for learning from noisy labels use static input features for denoising, but these methods are limited by the information they can provide on true label distributions and can result in biased or incorrect predictions.
no code implementations • 26 May 2023 • Gianfranco Cortes, Yue Yu, Robin Chen, Melissa Armstrong, David Vaillancourt, Baba C. Vemuri
With the advent of group equivariant convolutions in deep networks literature, spherical CNNs with $\mathsf{SO}(3)$-equivariant layers have been developed to cope with data that are samples of signals on the sphere $S^2$.
1 code implementation • 18 May 2023 • Yue Yu, Yuchen Zhuang, Rongzhi Zhang, Yu Meng, Jiaming Shen, Chao Zhang
With the development of large language models (LLMs), zero-shot learning has attracted much attention for various NLP tasks.
Ranked #1 on Zero-Shot Text Classification on AG News
no code implementations • 11 May 2023 • Qinshuang Wei, Yue Yu, Ufuk Topcu
Urban air mobility (UAM) is an emerging concept in short-range aviation transportation, where the aircraft will take off, land, and charge their batteries at a set of vertistops, and travel only through a set of flight corridors connecting these vertistops.
1 code implementation • 31 Mar 2023 • Shenghui Chen, Yue Yu, David Fridovich-Keil, Ufuk Topcu
Markov games model interactions among multiple players in a stochastic, dynamic environment.
no code implementations • 2 Mar 2023 • Dun Zeng, Xiangjing Hu, Shiyu Liu, Yue Yu, Qifan Wang, Zenglin Xu
Federated learning is a distributed learning framework that takes full advantage of private data samples kept on edge devices.
no code implementations • 28 Jan 2023 • Lu Zhang, Huaiqian You, Tian Gao, Mo Yu, Chung-Hao Lee, Yue Yu
Gradient-based meta-learning methods have primarily been applied to classical machine learning tasks such as image classification.
no code implementations • 11 Jan 2023 • Huaiqian You, Xiao Xu, Yue Yu, Stewart Silling, Marta D'Elia, John Foster
Then, based on the coarse-grained MD data, a two-phase optimization-based learning approach is proposed to infer the optimal peridynamics model with damage criterion.
1 code implementation • 10 Jan 2023 • ran Xu, Yue Yu, Hejie Cui, Xuan Kan, Yanqiao Zhu, Joyce Ho, Chao Zhang, Carl Yang
Our further analysis demonstrates that our proposed data selection strategy reduces the noise of pseudo labels by 36. 8% and saves 57. 3% of the time when compared with the best baseline.
no code implementations • 29 Dec 2022 • Ning Liu, Yue Yu, Huaiqian You, Neeraj Tatikola
Neural operators, which emerge as implicit solution operators of hidden governing equations, have recently become popular tools for learning responses of complex real-world physical systems.
no code implementations • 27 Dec 2022 • Xiaojun Xu, Yue Yu, Hanzhang Wang, Alok Lal, Carl A. Gunter, Bo Li
In this paper, we propose a general adversarial edge detection pipeline EDoG without requiring knowledge of the attack strategies based on graph generation.
no code implementations • 29 Nov 2022 • Huiyan Qi, Lechao Cheng, Jingjing Chen, Yue Yu, Xue Song, Zunlei Feng, Yu-Gang Jiang
Transfer learning aims to improve the performance of target tasks by transferring knowledge acquired in source tasks.
1 code implementation • 9 Nov 2022 • Mo Yu, Qiujing Wang, Shunchi Zhang, Yisi Sang, Kangsheng Pu, Zekai Wei, Han Wang, Liyan Xu, Jing Li, Yue Yu, Jie zhou
Our dataset consists of ~1, 000 parsed movie scripts, each corresponding to a few-shot character understanding task that requires models to mimic humans' ability of fast digesting characters with a few starting scenes in a new movie.
1 code implementation • 1 Nov 2022 • Yue Yu, Xuan Kan, Hejie Cui, ran Xu, Yujia Zheng, Xiangchen Song, Yanqiao Zhu, Kun Zhang, Razieh Nabi, Ying Guo, Chao Zhang, Carl Yang
To better adapt GNNs for fMRI analysis, we propose TBDS, an end-to-end framework based on \underline{T}ask-aware \underline{B}rain connectivity \underline{D}AG (short for Directed Acyclic Graph) \underline{S}tructure generation for fMRI analysis.
1 code implementation • 31 Oct 2022 • Si Sun, Chenyan Xiong, Yue Yu, Arnold Overwijk, Zhiyuan Liu, Jie Bao
In this paper, we investigate the instability in the standard dense retrieval training, which iterates between model training and hard negative selection using the being-trained model.
1 code implementation • 27 Oct 2022 • Yue Yu, Chenyan Xiong, Si Sun, Chao Zhang, Arnold Overwijk
We present a new zero-shot dense retrieval (ZeroDR) method, COCO-DR, to improve the generalization ability of dense retrieval by combating the distribution shifts between source training tasks and target scenarios.
Ranked #1 on Zero-shot Text Search on CQADupStack
1 code implementation • 26 Oct 2022 • Yuchen Zhuang, Yinghao Li, Jerry Junyang Cheung, Yue Yu, Yingjun Mou, Xiang Chen, Le Song, Chao Zhang
We study the problem of extracting N-ary relation tuples from scientific articles.
no code implementations • 19 Sep 2022 • Yue Yu, Ruihan Zhao, Sandeep Chinchali, Ufuk Topcu
Data-driven predictive control (DPC) is a feedback control method for systems with unknown dynamics.
1 code implementation • 15 Sep 2022 • Yue Yu, Rongzhi Zhang, ran Xu, Jieyu Zhang, Jiaming Shen, Chao Zhang
Large Language Models have demonstrated remarkable few-shot performance, but the performance can be sensitive to the selection of few-shot instances.
no code implementations • 11 Aug 2022 • Jia-Xin Zhuang, Xiansong Huang, Yang Yang, Jiancong Chen, Yue Yu, Wei Gao, Ge Li, Jie Chen, Tong Zhang
In this paper, we present OpenMedIA, an open-source toolbox library containing a rich set of deep learning methods for medical image analysis under heterogeneous Artificial Intelligence (AI) computing platforms.
1 code implementation • 8 Jul 2022 • Somdatta Goswami, Aniruddha Bora, Yue Yu, George Em Karniadakis
Standard neural networks can approximate general nonlinear operators, represented either explicitly by a combination of mathematical operators, e. g., in an advection-diffusion-reaction partial differential equation, or simply as a black box, e. g., a system-of-systems.
1 code implementation • 15 Jun 2022 • Wenzhong Shi, Pengxin Chen, Muyang Wang, Sheng Bao, Haodong Xiang, Yue Yu, Daping Yang
Color checker boards are pasted in each surveyed area as targets and ground truth data are collected by an advanced terrestrial laser scanner (TLS).
no code implementations • 4 Jun 2022 • Lu Zhang, Huaiqian You, Yue Yu
We propose MetaNOR, a meta-learnt approach for transfer-learning operators based on the nonlocal operator regression.
no code implementations • 23 May 2022 • Fei Lu, Qingci An, Yue Yu
In this work, we provide a rigorous identifiability analysis and convergence study for the learning of kernels in nonlocal operators.
1 code implementation • 19 May 2022 • Yang Xiang, Zhihua Wu, Weibao Gong, Siyu Ding, Xianjie Mo, Yuang Liu, Shuohuan Wang, Peng Liu, Yongshuai Hou, Long Li, Bin Wang, Shaohuai Shi, Yaqian Han, Yue Yu, Ge Li, Yu Sun, Yanjun Ma, dianhai yu
We took natural language processing (NLP) as an example to show how Nebula-I works in different training phases that include: a) pre-training a multilingual language model using two remote clusters; and b) fine-tuning a machine translation model using knowledge distilled from pre-trained models, which run through the most popular paradigm of recent deep learning.
no code implementations • 18 Apr 2022 • Xun Wang, Bingqing Ke, Xuanping Li, Fangyu Liu, Mingyu Zhang, Xiao Liang, Qiushi Xiao, Cheng Luo, Yue Yu
This modality imbalanceresults from a) modality gap: the relevance between a query and a video text is much easier to learn as the query is also a piece of text, with the same modality as the video text; b) data bias: most training samples can be solved solely by text matching.
1 code implementation • 1 Apr 2022 • Huaiqian You, Quinn Zhang, Colton J. Ross, Chung-Hao Lee, Ming-Chen Hsu, Yue Yu
To improve the generalizability of our framework, we propose a physics-guided neural operator learning model via imposing partial physics knowledge.
1 code implementation • 18 Mar 2022 • Rongzhi Zhang, Yue Yu, Pranav Shetty, Le Song, Chao Zhang
Weakly-supervised learning (WSL) has shown promising results in addressing label scarcity on many NLP tasks, but manually designing a comprehensive, high-quality labeling rule set is tedious and difficult.
1 code implementation • 15 Mar 2022 • Huaiqian You, Quinn Zhang, Colton J. Ross, Chung-Hao Lee, Yue Yu
In this work, we propose to use data-driven modeling, which directly utilizes high-fidelity simulation and/or experimental measurements to predict a material's response without using conventional constitutive models.
no code implementations • 2 Mar 2022 • Nan Xu, Jingchen Li, Yue Yu, Yang Li, Jinglei Yang
Positive feedback has been collected from pilot tests in several labs.
no code implementations • 25 Feb 2022 • Minglang Yin, Enrui Zhang, Yue Yu, George Em Karniadakis
In this work, we explore the idea of multiscale modeling with machine learning and employ DeepONet, a neural operator, as an efficient surrogate of the expensive solver.
1 code implementation • 12 Feb 2022 • Haibo Jin, Ruoxi Chen, Haibin Zheng, Jinyin Chen, Yao Cheng, Yue Yu, Xianglong Liu
By maximizing the number of excitable neurons concerning various wrong behaviors of models, DeepSensor can generate testing examples that effectively trigger more errors due to adversarial inputs, polluted data and incomplete training.
1 code implementation • 11 Feb 2022 • Jieyu Zhang, Cheng-Yu Hsieh, Yue Yu, Chao Zhang, Alexander Ratner
Labeling training data has become one of the major roadblocks to using machine learning.
2 code implementations • 12 Jan 2022 • Xiaoqi Wang, Yingjie Cheng, Yaning Yang, Yue Yu, Fei Li, Shaoliang Peng
Therefore, we conjecture that the multimodal and local-global combination strategies can be treated as the guideline of multi-task SSL for drug discovery.
no code implementations • 6 Jan 2022 • Huaiqian You, Yue Yu, Marta D'Elia, Tian Gao, Stewart Silling
In this work, we propose a novel nonlocal neural operator, which we refer to as nonlocal kernel network (NKN), that is resolution independent, characterized by deep neural networks, and capable of handling a variety of tasks such as learning governing equations and classifying images.
1 code implementation • 25 Dec 2021 • Haibin Zheng, Zhiqing Chen, Tianyu Du, Xuhong Zhang, Yao Cheng, Shouling Ji, Jingyi Wang, Yue Yu, Jinyin Chen
To overcome the challenges, we propose NeuronFair, a new DNN fairness testing framework that differs from previous work in several key aspects: (1) interpretable - it quantitatively interprets DNNs' fairness violations for the biased decision; (2) effective - it uses the interpretation results to guide the generation of more diverse instances in less time; (3) generic - it can handle both structured and unstructured data.
1 code implementation • 16 Dec 2021 • Yue Yu, Lingkai Kong, Jieyu Zhang, Rongzhi Zhang, Chao Zhang
We propose {\ours}, a new framework that leverages unlabeled data to improve the label efficiency of active PLM fine-tuning.
1 code implementation • 4 Dec 2021 • Deze Wang, Zhouyang Jia, Shanshan Li, Yue Yu, Yun Xiong, Wei Dong, Xiangke Liao
In this paper, we propose an approach to bridge pre-trained models and code-related tasks.
no code implementations • 16 Oct 2021 • Zhijing Zhang, Yue Yu, Qinghua Ma, Haixiang Yao
In allusion to some contradicting results in existing research, this paper selects China's latest stock data from 2005 to 2020 for empirical analysis.
1 code implementation • 23 Sep 2021 • Jieyu Zhang, Yue Yu, Yinghao Li, Yujing Wang, Yaming Yang, Mao Yang, Alexander Ratner
To address these problems, we introduce a benchmark platform, WRENCH, for thorough and standardized evaluation of WS approaches.
no code implementations • Findings (NAACL) 2022 • Simiao Zuo, Yue Yu, Chen Liang, Haoming Jiang, Siawpeng Er, Chao Zhang, Tuo Zhao, Hongyuan Zha
In self-training, the student contributes to the prediction performance, and the teacher controls the training process by generating pseudo-labels.
no code implementations • 16 Aug 2021 • Somdatta Goswami, Minglang Yin, Yue Yu, George Karniadakis
We propose a physics-informed variational formulation of DeepONet (V-DeepONet) for brittle fracture analysis.
1 code implementation • 5 Aug 2021 • Danylo Malyuta, Yue Yu, Purnanand Elango, Behcet Acikmese
Space mission design places a premium on cost and operational efficiency.
no code implementations • 4 Aug 2021 • Huaiqian You, Yue Yu, Stewart Silling, Marta D'Elia
Nonlocal models, including peridynamics, often use integral operators that embed lengthscales in their definition.
no code implementations • 9 Jul 2021 • Zuohui Chen, Renxuan Wang, Jingyang Xiang, Yue Yu, Xin Xia, Shouling Ji, Qi Xuan, Xiaoniu Yang
Deep Neural Networks (DNN) are known to be vulnerable to adversarial samples, the detection of which is crucial for the wide application of these DNN models.
1 code implementation • 14 Jun 2021 • Yue Yu, Tian Gao, Naiyu Yin, Qiang Ji
To further improve efficiency, we propose a novel learning framework to model and learn the weighted adjacency matrices in the DAG space directly.
4 code implementations • 26 Apr 2021 • Wei Zeng, Xiaozhe Ren, Teng Su, Hui Wang, Yi Liao, Zhiwei Wang, Xin Jiang, ZhenZhang Yang, Kaisheng Wang, Xiaoda Zhang, Chen Li, Ziyan Gong, Yifan Yao, Xinjing Huang, Jun Wang, Jianfeng Yu, Qi Guo, Yue Yu, Yan Zhang, Jin Wang, Hengtao Tao, Dasen Yan, Zexuan Yi, Fang Peng, Fangqing Jiang, Han Zhang, Lingfeng Deng, Yehong Zhang, Zhe Lin, Chao Zhang, Shaojie Zhang, Mingyue Guo, Shanzhi Gu, Gaojun Fan, YaoWei Wang, Xuefeng Jin, Qun Liu, Yonghong Tian
To enhance the generalization ability of PanGu-$\alpha$, we collect 1. 1TB high-quality Chinese data from a wide range of domains to pretrain the model.
Ranked #1 on Reading Comprehension (One-Shot) on DuReader
Cloze (multi-choices) (Few-Shot) Cloze (multi-choices) (One-Shot) +19
1 code implementation • 24 Mar 2021 • Chen Zeng, Yue Yu, Shanshan Li, Xin Xia, Zhiming Wang, Mingyang Geng, Bailin Xiao, Wei Dong, Xiangke Liao
With the rapid increase in the amount of public code repositories, developers maintain a great desire to retrieve precise code snippets by using natural language.
no code implementations • 5 Feb 2021 • Yue Yu, Shahriar Talebi, Henk J. van Waarde, Ufuk Topcu, Mehran Mesbahi, Behçet Açıkmeşe
Willems' fundamental lemma asserts that all trajectories of a linear time-invariant system can be obtained from a finite number of measured ones, assuming that controllability and a persistency of excitation condition hold.
no code implementations • 5 Jan 2021 • Yue Yu, Huaiqian You, Nathaniel Trask
In the absence of fracture, when a corresponding classical continuum mechanics model exists, our improvements provide asymptotically compatible convergence to corresponding local solutions, eliminating surface effects and issues with traction loading which have historically plagued peridynamic discretizations.
Numerical Analysis Computational Engineering, Finance, and Science Numerical Analysis Analysis of PDEs
no code implementations • 8 Dec 2020 • Huaiqian You, Yue Yu, Stewart Silling, Marta D'Elia
We show that machine learning can improve the accuracy of simulations of stress waves in one-dimensional composite materials.
1 code implementation • NeurIPS 2020 • Dennis Wei, Tian Gao, Yue Yu
This paper re-examines a continuous optimization framework dubbed NOTEARS for learning Bayesian networks.
1 code implementation • NAACL 2021 • Yue Yu, Simiao Zuo, Haoming Jiang, Wendi Ren, Tuo Zhao, Chao Zhang
To address this problem, we develop a contrastive self-training framework, COSINE, to enable fine-tuning LMs with weak supervision.
Ranked #1 on Word Sense Disambiguation on Words in Context
1 code implementation • EMNLP 2020 • Rongzhi Zhang, Yue Yu, Chao Zhang
Our method, SeqMix, simply augments the queried samples by generating extra labeled sequences in each iteration.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Lu Zhang, Mo Yu, Tian Gao, Yue Yu
Multi-hop reasoning approaches over knowledge graphs infer a missing relationship between entities with a multi-hop rule, which corresponds to a chain of relationships.
1 code implementation • 4 Oct 2020 • Yue Yu, Kexin Huang, Chao Zhang, Lucas M. Glass, Jimeng Sun, Cao Xiao
Furthermore, most previous works focus on binary DDI prediction whereas the multi-typed DDI pharmacological effect prediction is a more meaningful but harder task.
1 code implementation • 28 Jun 2020 • Chen Liang, Yue Yu, Haoming Jiang, Siawpeng Er, Ruijia Wang, Tuo Zhao, Chao Zhang
We study the open-domain named entity recognition (NER) problem under distant supervision.
1 code implementation • 18 Jun 2020 • Yue Yu, Yinghao Li, Jiaming Shen, Hao Feng, Jimeng Sun, Chao Zhang
We propose a self-supervised taxonomy expansion model named STEAM, which leverages natural supervision in the existing taxonomy for expansion.
no code implementations • 17 May 2020 • Huaiqian You, Yue Yu, Nathaniel Trask, Mamikon Gulian, Marta D'Elia
A key challenge to nonlocal models is the analytical complexity of deriving them from first principles, and frequently their use is justified a posteriori.
no code implementations • 25 Apr 2020 • Mingyang Zhang, Tong Li, Yue Yu, Yong Li, Pan Hui, Yu Zheng
Urban anomalies may result in loss of life or property if not handled properly.
no code implementations • 10 Feb 2020 • Xunpeng Huang, Xianfeng Liang, Zhengyang Liu, Yitan Li, Linyun Yu, Yue Yu, Lei LI
SPAN computes the inverse of the Hessian matrix via low-rank approximation and stochastic Hessian-vector products.
no code implementations • 25 Sep 2019 • Xunpeng Huang, Zhengyang Liu, Zhe Wang, Yue Yu, Lei LI
To the best of our knowledge, Acutum is the first adaptive gradient method without second moments.
no code implementations • 2 Sep 2019 • Yue Yu, Siyao Peng, Grace Hui Yang
Previous work on DA recognition either assumes one DA per utterance or fails to realize the sequential nature of dialogues.
1 code implementation • WS 2019 • Yue Yu, YIlun Zhu, Yang Liu, Yan Liu, Siyao Peng, Mackenzie Gong, Amir Zeldes
In this paper we present GumDrop, Georgetown University's entry at the DISRPT 2019 Shared Task on automatic discourse unit segmentation and connective detection.
3 code implementations • 22 Apr 2019 • Yue Yu, Jie Chen, Tian Gao, Mo Yu
Learning a faithful directed acyclic graph (DAG) from samples of a joint distribution is a challenging combinatorial problem, owing to the intractable search space superexponential in the number of graph nodes.
no code implementations • 27 Sep 2018 • Xiaojun Xu, Yue Yu, Bo Li, Le Song, Chengfeng Liu, Carl Gunter
Extensive experiments are conducted to show that the proposed detection mechanism can achieve AUC above 90% against the two attack strategies on both Cora and Citeseer datasets.
no code implementations • NeurIPS 2019 • Yue Yu, Jiaxiang Wu, Longbo Huang
In this paper, to reduce the communication complexity, we propose \emph{double quantization}, a general scheme for quantizing both model parameters and gradients.
1 code implementation • 10 Feb 2018 • Hongliang Liang, Mingyu Li, Qiong Zhang, Yue Yu, Lin Jiang, Yixiu Chen
Intel SGX provisions shielded executions for security-sensitive computation, but lacks support for trusted system services (TSS), such as clock, network and filesystem.
Cryptography and Security
no code implementations • 24 May 2017 • Scott Cheng-Hsin Yang, Yue Yu, Arash Givchi, Pei Wang, Wai Keen Vong, Patrick Shafto
Cooperative transmission of data fosters rapid accumulation of knowledge by efficiently combining experiences across learners.
no code implementations • 11 May 2017 • Yue Yu, Longbo Huang
We consider the stochastic composition optimization problem proposed in \cite{wang2017stochastic}, which has applications ranging from estimation to statistical and machine learning.
no code implementations • LREC 2012 • Ying Li, Yue Yu, Pascale Fung
Generally the existing monolingual corpora are not suitable for large vocabulary continuous speech recognition (LVCSR) of code-switching speech.