1 code implementation • ACL (dialdoc) 2021 • Liu Yang, Fanqi Meng, Xiao Liu, Ming-Kuang Daniel Wu, Vicent Ying, James Xu
In this work, we formulate a visual dialog as an information flow in which each piece of information is encoded with the joint visual-linguistic representation of a single dialog round.
no code implementations • LREC 2022 • Liu Yang, Catherine Achard, Catherine Pelachaud
Integrating the existing interruption and turn switch classification methods, we propose a new annotation schema to annotate different types of interruptions through timeliness, switch accomplishment and speech content level.
no code implementations • 19 Feb 2025 • Feiyuan Zhang, Dezhi Zhu, James Ming, Yilun Jin, Di Chai, Liu Yang, Han Tian, Zhaoxin Fan, Kai Chen
Retrieval-Augmented Generation (RAG) systems have shown substantial benefits in applications such as question answering and multi-turn dialogue \citep{lewis2020retrieval}.
1 code implementation • 9 Feb 2025 • Elisa Negrini, Yuxuan Liu, Liu Yang, Stanley J. Osher, Hayden Schaeffer
PDE foundation models utilize neural networks to train approximations to multiple differential equations simultaneously and are thus a general purpose solver that can be adapted to downstream tasks.
no code implementations • 16 Jan 2025 • Liu Yang, Ziqian Lin, Kangwook Lee, Dimitris Papailiopoulos, Robert Nowak
In-context learning is a remarkable capability of transformers, referring to their ability to adapt to specific tasks based on a short history or context.
no code implementations • 2 Jan 2025 • Zitong Xu, Huiyu Duan, Guangji Ma, Liu Yang, Jiarui Wang, Qingbo Wu, Xiongkuo Min, Guangtao Zhai, Patrick Le Callet
To address the issue and facilitate the advancement of IHAs, we introduce the first Image Quality Assessment Database for image Harmony evaluation (HarmonyIQAD), which consists of 1, 350 harmonized images generated by 9 different IHAs, and the corresponding human visual preference scores.
no code implementations • 29 Dec 2024 • Xilei Zhu, Huiyu Duan, Liu Yang, Yucheng Zhu, Xiongkuo Min, Guangtao Zhai, Patrick Le Callet
With the rapid development of eXtended Reality (XR), egocentric spatial shooting and display technologies have further enhanced immersion and engagement for users.
no code implementations • 27 Dec 2024 • Wang Qun, Liu Yang, Lin Qingquan, Qu Zhijiu, Jiang Ling
Xmodel-2 is a 1. 2-billion-parameter large language model designed specifically for reasoning tasks.
no code implementations • 26 Dec 2024 • Huiyu Duan, Qiang Hu, Jiarui Wang, Liu Yang, Zitong Xu, Lu Liu, Xiongkuo Min, Chunlei Cai, Tianxiao Ye, Xiaoyun Zhang, Guangtao Zhai
The rapid growth of user-generated content (UGC) videos has produced an urgent need for effective video quality assessment (VQA) algorithms to monitor video quality and guide optimization and recommendation procedures.
no code implementations • 23 Dec 2024 • Xiangfei Qiu, Xiuwen Li, Ruiyang Pang, Zhicheng Pan, Xingjian Wu, Liu Yang, Jilin Hu, Yang Shu, Xuesong Lu, Chengcheng Yang, Chenjuan Guo, Aoying Zhou, Christian S. Jensen, Bin Yang
First, EasyTime enables one-click evaluation, enabling researchers to evaluate new forecasting methods using the suite of diverse time series datasets collected in the preexisting time series forecasting benchmark (TFB).
no code implementations • 17 Dec 2024 • Lu Liu, Huiyu Duan, Qiang Hu, Liu Yang, Chunlei Cai, Tianxiao Ye, Huayu Liu, Xiaoyun Zhang, Guangtao Zhai
The FaceQ database comprises 12, 255 images generated by 29 models across three tasks: (1) face generation, (2) face customization, and (3) face restoration.
no code implementations • 11 Dec 2024 • Fabian Paischer, Liu Yang, Linfeng Liu, Shuai Shao, Kaveh Hassani, Jiacheng Li, Ricky Chen, Zhang Gabriel Li, Xialo Gao, Wei Shao, Xue Feng, Nima Noorshams, Sem Park, Bo Long, Hamid Eghbalzadeh
We assess current state-of-the-art methods using our benchmark and show that they struggle to accurately discern user preferences.
1 code implementation • 8 Dec 2024 • Ma Teng, Jia Xiaojun, Duan Ranjie, Li Xinfeng, Huang Yihao, Chu Zhixuan, Liu Yang, Ren Wenqi
The multimodal risk distribution strategy is used to segment harmful instructions across multiple modalities to effectively circumvent MLLMs' security protection.
no code implementations • 27 Nov 2024 • Liu Yang, Fabian Paischer, Kaveh Hassani, Jiacheng Li, Shuai Shao, Zhang Gabriel Li, Yun He, Xue Feng, Nima Noorshams, Sem Park, Bo Long, Robert D Nowak, Xiaoli Gao, Hamid Eghbalzadeh
This hybrid approach provides insights into the trade-offs between these approaches and demonstrates improvements in efficiency and effectiveness for recommendation systems in small-scale benchmarks.
no code implementations • 25 Nov 2024 • Yadi Cao, Yuxuan Liu, Liu Yang, Rose Yu, Hayden Schaeffer, Stanley Osher
In-Context Operator Networks (ICONs) are models that learn operators across different types of PDEs using a few-shot, in-context approach.
1 code implementation • 15 Nov 2024 • Wang Qun, Liu Yang, Lin Qingquan, Jiang Ling
We introduce Xmodel-1. 5, a 1-billion-parameter multilingual large language model pretrained on 2 trillion tokens, designed for balanced performance and scalability.
no code implementations • 6 Nov 2024 • Ying Zhang, Qiang Li, Hongli Liu, Liu Yang, Jian Yang
Radio Frequency Fingerprint Identification (RFFI) technology uniquely identifies emitters by analyzing unique distortions in the transmitted signal caused by non-ideal hardware.
no code implementations • 3 Nov 2024 • Yun He, Xuxing Chen, Jiayi Xu, Renqin Cai, Yiling You, Jennifer Cao, Minhui Huang, Liu Yang, Yiqun Liu, Xiaoyi Liu, Rong Jin, Sem Park, Bo Long, Xue Feng
In industrial recommendation systems, multi-task learning (learning multiple tasks simultaneously on a single model) is a predominant approach to save training/serving resources and improve recommendation performance via knowledge transfer between the joint learning tasks.
no code implementations • 8 Oct 2024 • Zheyang Xiong, Ziyang Cai, John Cooper, Albert Ge, Vasilis Papageorgiou, Zack Sifakis, Angeliki Giannou, Ziqian Lin, Liu Yang, Saurabh Agarwal, Grigorios G Chrysos, Samet Oymak, Kangwook Lee, Dimitris Papailiopoulos
In this study, we explore a surprising phenomenon related to ICL: LLMs can perform multiple, computationally distinct ICL tasks simultaneously, during a single inference call, a capability we term "task superposition".
no code implementations • 31 Jul 2024 • Xilei Zhu, Liu Yang, Huiyu Duan, Xiongkuo Min, Guangtao Zhai, Patrick Le Callet
In this paper, we establish the Egocentric Spatial Images Quality Assessment Database (ESIQAD), the first IQA database dedicated for egocentric spatial images as far as we know.
no code implementations • 11 Jul 2024 • Liang Zeng, Liangjun Zhong, Liang Zhao, Tianwen Wei, Liu Yang, Jujie He, Cheng Cheng, Rui Hu, Yang Liu, Shuicheng Yan, Han Fang, Yahui Zhou
In this paper, we investigate the underlying factors that potentially enhance the mathematical reasoning capabilities of large language models (LLMs).
no code implementations • 2 Jun 2024 • Liang Zhao, Tianwen Wei, Liang Zeng, Cheng Cheng, Liu Yang, Peng Cheng, Lijie Wang, Chenxia Li, Xuejie Wu, Bo Zhu, Yimeng Gan, Rui Hu, Shuicheng Yan, Han Fang, Yahui Zhou
We introduce LongSkywork, a long-context Large Language Model (LLM) capable of processing up to 200, 000 tokens.
no code implementations • 1 May 2024 • Liu Yang, Shuowei Cai, Di Chai, Junxue Zhang, Han Tian, Yilun Jin, Kun Guo, Kai Chen, Qiang Yang
To this core, we propose PackVFL, an efficient VFL framework based on packed HE (PackedHE), to accelerate the existing HE-based VFL algorithms.
no code implementations • 16 Apr 2024 • Ziqi Zhao, Zhaochun Ren, Liu Yang, Fajie Yuan, Pengjie Ren, Zhumin Chen, Jun Ma, Xin Xin
Then we propose four strategies to use World Transformers to generate high-rewarded trajectory simulation by perturbing the offline data.
no code implementations • 12 Apr 2024 • Liu Yang, Qiang Li, Xiaoyang Ren, Yi Fang, Shafei Wang
To address this issue, we formulate the cross-receiver RFFI as a model adaptation problem, which adapts the trained model to unlabeled signals from a new receiver.
1 code implementation • 1 Apr 2024 • Liu Yang, Huiyu Duan, Long Teng, Yucheng Zhu, Xiaohong Liu, Menghan Hu, Xiongkuo Min, Guangtao Zhai, Patrick Le Callet
Finally, we conduct a benchmark experiment to evaluate the performance of state-of-the-art IQA models on our database.
no code implementations • 15 Mar 2024 • Changhong Hou, Junchuan Yu, Daqing Ge, Liu Yang, Laidian Xi, Yunxuan Pang, Yi Wen
We propose TransLandSeg, which is a transfer learning approach for landslide semantic segmentation based on a vision foundation model (VFM).
no code implementations • 5 Mar 2024 • Angeliki Giannou, Liu Yang, Tianhao Wang, Dimitris Papailiopoulos, Jason D. Lee
Recent studies have suggested that Transformers can implement first-order optimization algorithms for in-context learning and even second order ones for the case of linear regression.
no code implementations • 15 Feb 2024 • Saeed Khaki, Jinjin Li, Lan Ma, Liu Yang, Prathap Ramachandra
Finally, we apply DPO with the contrastive samples to align the model to human preference.
no code implementations • 11 Feb 2024 • Jeongyeol Kwon, Liu Yang, Robert Nowak, Josiah Hanna
Then, our main contributions are two-fold: (a) we demonstrate that the performance of reinforcement learning is strongly correlated with the prediction accuracy of future observations in partially observable environments, and (b) our approach can significantly improve the overall end-to-end approach by preventing high-variance noisy signals from reinforcement learning objectives to influence the representation learning.
1 code implementation • 22 Jan 2024 • Haiqian Yang, Florian Meyer, Shaoxun Huang, Liu Yang, Cristiana Lungu, Monilola A. Olayioye, Markus J. Buehler, Ming Guo
Multicellular self-assembly into functional structures is a dynamic process that is critical in the development and diseases, including embryo development, organ formation, tumor invasion, and others.
1 code implementation • 14 Jan 2024 • Liu Yang, Stanley J. Osher
We show the positive evidence to the second question, i. e., ICON can generalize well to some PDEs with new forms without any fine-tuning.
1 code implementation • 22 Dec 2023 • Xin Xin, Liu Yang, Ziqi Zhao, Pengjie Ren, Zhumin Chen, Jun Ma, Zhaochun Ren
On the one hand, these approaches cannot achieve satisfying unlearning effects due to the collaborative correlations and sequential connections between the unlearning item and the remaining items in the session.
no code implementations • 27 Nov 2023 • Can Sun, Hao Zheng, Zhigang Hu, Liu Yang, Meiguang Zheng, Bo Xu
The single domain generalization(SDG) based on meta-learning has emerged as an effective technique for solving the domain-shift problem.
no code implementations • 26 Nov 2023 • Xianting Feng, Hao Zheng, Zhigang Hu, Liu Yang, Meiguang Zheng
Most existing synthetic aperture radar (SAR) ship classification technologies heavily rely on correctly labeled data, ignoring the discriminative features of unlabeled SAR ship images.
no code implementations • 26 Nov 2023 • Bo Xu, Hao Zheng, Zhigang Hu, Liu Yang, Meiguang Zheng
In current synthetic aperture radar (SAR) object classification, one of the major challenges is the severe overfitting issue due to the limited dataset (few-shot) and noisy data.
1 code implementation • 21 Nov 2023 • Liu Yang, Kangwook Lee, Robert Nowak, Dimitris Papailiopoulos
Transformers have demonstrated effectiveness in in-context solving data-fitting problems from various (latent) models, as reported by Garg et al.
1 code implementation • 30 Oct 2023 • Tianwen Wei, Liang Zhao, Lichang Zhang, Bo Zhu, Lijie Wang, Haihua Yang, Biye Li, Cheng Cheng, Weiwei Lü, Rui Hu, Chenxia Li, Liu Yang, Xilin Luo, Xuejie Wu, Lunan Liu, Wenjun Cheng, Peng Cheng, Jianhao Zhang, XiaoYu Zhang, Lei Lin, Xiaokun Wang, Yutuan Ma, Chuanhai Dong, Yanqi Sun, Yifu Chen, Yongyi Peng, Xiaojuan Liang, Shuicheng Yan, Han Fang, Yahui Zhou
In this technical report, we present Skywork-13B, a family of large language models (LLMs) trained on a corpus of over 3. 2 trillion tokens drawn from both English and Chinese texts.
1 code implementation • 25 Oct 2023 • Liu Yang, Haihua Yang, Wenjun Cheng, Lei Lin, Chenxia Li, Yifu Chen, Lunan Liu, Jianfei Pan, Tianwen Wei, Biye Li, Liang Zhao, Lijie Wang, Bo Zhu, Guoliang Li, Xuejie Wu, Xilin Luo, Rui Hu
Large language models (LLMs) have shown great potential to solve varieties of natural language processing (NLP) tasks, including mathematical reasoning.
1 code implementation • 23 Aug 2023 • Di Chai, Leye Wang, Liu Yang, Junxue Zhang, Kai Chen, Qiang Yang
Evaluation is a systematic approach to assessing how well a system achieves its intended purpose.
1 code implementation • 9 Aug 2023 • Liu Yang, Siting Liu, Stanley J. Osher
In the growing domain of scientific machine learning, in-context operator learning has shown notable potential in building foundation models, as in this framework the model is trained to learn operators and solve differential equations using prompted data, during the inference stage without weight updates.
2 code implementations • 17 Apr 2023 • Liu Yang, Siting Liu, Tingwei Meng, Stanley J. Osher
This paper introduces a new neural-network-based approach, namely In-Context Operator Networks (ICON), to simultaneously learn operators from the prompted data and apply it to new questions during the inference stage, without any weight update.
no code implementations • 5 Apr 2023 • Qihang Zhao, Rui-Jie Zhu, Liu Yang, He Yongming, Bo Zhou, Luo Cheng
In the realm of search systems, multi-stage cascade architecture is a prevalent method, typically consisting of sequential modules such as matching, pre-ranking, and ranking.
no code implementations • 4 Apr 2023 • Liu Yang, Di Chai, Junxue Zhang, Yilun Jin, Leye Wang, Hao liu, Han Tian, Qian Xu, Kai Chen
From the hardware layer to the vertical federated system layer, researchers contribute to various aspects of VFL.
no code implementations • 23 Jan 2023 • Kunlong Chen, Liu Yang, Yitian Chen, Kunjin Chen, Yidan Xu, Lujun Li
It is of great significance to estimate the performance of a given model architecture without training in the application of Neural Architecture Search (NAS) as it may take a lot of time to evaluate the performance of an architecture.
no code implementations • CVPR 2023 • Zixuan Qin, Liu Yang, Qilong Wang, Yahong Han, QinGhua Hu
When there are large differences in data distribution among clients, it is crucial for federated learning to design a reliable client selection strategy and an interpretable client communication framework to better utilize group knowledge.
no code implementations • 14 Oct 2022 • Liu Yang, Simon X. Yang, Yun Li, Yinzhi Lu, Tan Guo
Traditional machine learning techniques have been widely used to establish the trust management systems.
no code implementations • 6 Oct 2022 • Liu Yang, Jifan Zhang, Joseph Shenouda, Dimitris Papailiopoulos, Kangwook Lee, Robert D. Nowak
Weight decay is one of the most widely used forms of regularization in deep learning, and has been shown to improve generalization and robustness.
no code implementations • 2 Aug 2022 • Liu Yang, Yun Li, Simon X. Yang, Yinzhi Lu, Tan Guo, Keping Yu
Next, the integration of AI and trust management is developed to optimize the intelligence and security.
no code implementations • 21 Jul 2022 • Liu Yang, Yinzhi Lu, Simon X. Yang, Yuanchang Zhong, Tan Guo, Zhifang Liang
To acquire secured data delivery and address the conflict between security and energy, in this paper we present an evolutionary game based secure clustering protocol with fuzzy trust evaluation and outlier detection for WSNs.
no code implementations • 20 Jul 2022 • Liu Yang, Yinzhi Lu, Simon X. Yang, Tan Guo, Zhifang Liang
And then a density based outlier detection mechanism is introduced to acquire an adaptive trust threshold used to isolate the malicious nodes from being cluster heads.
no code implementations • 19 Jul 2022 • Liu Yang, Keping Yu, Simon X. Yang, Chinmay Chakraborty, Yinzhi Lu, Tan Guo
To assure secure and reliable communication in 5G edge computing and D2D enabled IoMT systems, this paper presents an intelligent trust cloud management method.
no code implementations • 17 Jul 2022 • Yinzhi Lu, Liu Yang, Simon X. Yang, Qiaozhi Hua, Arun Kumar Sangaiah, Tan Guo, Keping Yu
Then a non-collision theory based deterministic scheduling (NDS) method is proposed to achieve ultra-low latency communication for the time-sensitive flows.
no code implementations • 28 Jun 2022 • Shuowei Cai, Di Chai, Liu Yang, Junxue Zhang, Yilun Jin, Leye Wang, Kun Guo, Kai Chen
In this paper, we focus on SplitNN, a well-known neural network framework in VFL, and identify a trade-off between data security and model performance in SplitNN.
1 code implementation • 24 Feb 2022 • Kartik Sreenivasan, Jy-yong Sohn, Liu Yang, Matthew Grinde, Alliot Nagle, Hongyi Wang, Eric Xing, Kangwook Lee, Dimitris Papailiopoulos
Frankle & Carbin conjecture that we can avoid this by training "lottery tickets", i. e., special sparse subnetworks found at initialization, that can be trained to high accuracy.
no code implementations • 18 Aug 2021 • Liu Yang, Junxue Zhang, Di Chai, Leye Wang, Kun Guo, Kai Chen, Qiang Yang
In this paper, we proposed federated masked matrix factorization (FedMMF) to protect the data privacy in federated recommender systems without sacrificing efficiency and effectiveness.
no code implementations • 30 Jun 2021 • Yisroel Mirsky, Ambra Demontis, Jaidip Kotak, Ram Shankar, Deng Gelei, Liu Yang, Xiangyu Zhang, Wenke Lee, Yuval Elovici, Battista Biggio
Although offensive AI has been discussed in the past, there is a need to analyze and understand the threat in the context of organizations.
no code implementations • 8 Jun 2021 • Xuhui Meng, Liu Yang, Zhiping Mao, Jose del Aguila Ferrandis, George Em Karniadakis
In summary, the proposed method is capable of learning flexible functional priors, and can be extended to big data problems using stochastic HMC or normalizing flows since the latent space is generally characterized as low dimensional.
1 code implementation • 9 May 2021 • Chen Qu, Hamed Zamani, Liu Yang, W. Bruce Croft, Erik Learned-Miller
We first conduct sparse retrieval with BM25 and study expanding the question with object names and image captions.
no code implementations • 15 Apr 2021 • Chen Qu, Weize Kong, Liu Yang, Mingyang Zhang, Michael Bendersky, Marc Najork
We investigate the privacy and utility implications of applying dx-privacy, a variant of Local Differential Privacy, to BERT fine-tuning in NLU applications.
1 code implementation • 3 Mar 2021 • Chen Qu, Liu Yang, Cen Chen, W. Bruce Croft, Kalpesh Krishna, Mohit Iyyer
Our method is more flexible as it can handle both span answers and freeform answers.
no code implementations • 17 Jan 2021 • Liu Yang, Tingwei Meng, George Em Karniadakis
We propose a simple but effective modification of the discriminators, namely measure-conditional discriminators, as a plug-and-play module for different GANs.
no code implementations • ICLR 2021 • Yi Tay, Mostafa Dehghani, Samira Abnar, Yikang Shen, Dara Bahri, Philip Pham, Jinfeng Rao, Liu Yang, Sebastian Ruder, Donald Metzler
Transformers do not scale very well to long sequence lengths largely because of quadratic self-attention complexity.
1 code implementation • 18 Dec 2020 • Xingjian Zhen, Rudrasis Chakraborty, Liu Yang, Vikas Singh
Partly due to this gap, there are also no modality transfer/translation models for manifold-valued data whereas numerous such methods based on generative models are available for natural images.
no code implementations • 25 Nov 2020 • Haojie Pan, Cen Chen, Chengyu Wang, Minghui Qiu, Liu Yang, Feng Ji, Jun Huang
More specifically, we propose a reinforced selector to extract useful PRF terms to enhance response candidates and a BERT-based response ranker to rank the PRF-enhanced responses.
no code implementations • 19 Nov 2020 • Di Chai, Leye Wang, Liu Yang, Junxue Zhang, Kai Chen, Qiang Yang
In this paper, we propose a holistic evaluation framework for FL called FedEval, and present a benchmarking study on seven state-of-the-art FL algorithms.
5 code implementations • 8 Nov 2020 • Yi Tay, Mostafa Dehghani, Samira Abnar, Yikang Shen, Dara Bahri, Philip Pham, Jinfeng Rao, Liu Yang, Sebastian Ruder, Donald Metzler
In the recent months, a wide spectrum of efficient, fast Transformers have been proposed to tackle this problem, more often than not claiming superior or comparable model quality to vanilla Transformer models.
Ranked #4 on
ListOps
on ListOps
no code implementations • Findings of the Association for Computational Linguistics 2020 • Jiecao Chen, Liu Yang, Karthik Raman, Michael Bendersky, Jung-Jung Yeh, Yun Zhou, Marc Najork, Danyang Cai, Ehsan Emadzadeh
Pre-trained models like BERT (Devlin et al., 2018) have dominated NLP / IR applications such as single sentence classification, text pair classification, and question answering.
no code implementations • 24 Aug 2020 • Xiaoli Chen, Liu Yang, Jinqiao Duan, George Em. Karniadakis
The Fokker-Planck (FP) equation governing the evolution of the probability density function (PDF) is applicable to many disciplines but it requires specification of the coefficients for each case, which can be functions of space-time and not just constants, hence requiring the development of a data-driven modeling approach.
no code implementations • 5 Aug 2020 • Liu Yang, Constantinos Daskalakis, George Em. Karniadakis
Particle coordinates at a single time instant, possibly noisy or truncated, are recorded in each snapshot but are unpaired across the snapshots.
1 code implementation • 2 Aug 2020 • Liu Yang
IP based SeqDialN is our baseline with a simple 2-layer LSTM design that achieves decent performance.
no code implementations • 7 Jul 2020 • Weiyan Wang, Cengguang Zhang, Liu Yang, Kai Chen, Kun Tan
However, due to the global synchronization nature, its performance can be significantly influenced by network bottlenecks caused by either static topology heterogeneity or dynamic bandwidth contentions.
no code implementations • 21 Jun 2020 • Zizhen Wang, Yixing Fan, Jiafeng Guo, Liu Yang, Ruqing Zhang, Yanyan Lan, Xue-Qi Cheng, Hui Jiang, Xiaozhao Wang
However, it has long been a challenge to properly measure the similarity between two questions due to the inherent variation of natural language, i. e., there could be different ways to ask a same question or different questions sharing similar expressions.
1 code implementation • 22 May 2020 • Chen Qu, Liu Yang, Cen Chen, Minghui Qiu, W. Bruce Croft, Mohit Iyyer
We build an end-to-end system for ORConvQA, featuring a retriever, a reranker, and a reader that are all based on Transformers.
1 code implementation • 26 Apr 2020 • Liu Yang, Mingyang Zhang, Cheng Li, Michael Bendersky, Marc Najork
In order to better capture sentence level semantic relations within a document, we pre-train the model with a novel masked sentence block language modeling task in addition to the masked word language modeling task used by BERT.
no code implementations • 13 Mar 2020 • Liu Yang, Xuhui Meng, George Em. Karniadakis
In this Bayesian framework, the Bayesian neural network (BNN) combined with a PINN for PDEs serves as the prior while the Hamiltonian Monte Carlo (HMC) or the variational inference (VI) could serve as an estimator of the posterior.
1 code implementation • 6 Mar 2020 • Dixia Fan, Liu Yang, Michael S. Triantafyllou, George Em. Karniadakis
We demonstrate experimentally the feasibility of applying reinforcement learning (RL) in flow control problems by automatically discovering active control strategies without any prior knowledge of the flow physics.
Fluid Dynamics Robotics
1 code implementation • ICML 2020 • Yi Tay, Dara Bahri, Liu Yang, Donald Metzler, Da-Cheng Juan
We propose Sparse Sinkhorn Attention, a new efficient and sparse method for learning to attend.
no code implementations • 20 Feb 2020 • Kristiaan Pelckmans, Liu Yang
This paper considers the task of learning how to make a prognosis of a patient based on his/her micro-array expression levels.
1 code implementation • 3 Feb 2020 • Liu Yang, Minghui Qiu, Chen Qu, Cen Chen, Jiafeng Guo, Yongfeng Zhang, W. Bruce Croft, Haiqing Chen
We also perform case studies and analysis of learned user intent and its impact on response ranking in information-seeking conversations to provide interpretation of results.
no code implementations • 5 Nov 2019 • Liu Yang, Rudrasis Chakraborty
Though in the medical imaging community, 3D point-cloud processing is not a "go-to" choice, it is a canonical way to preserve rotation invariance.
no code implementations • 5 Nov 2019 • Liu Yang, Rudrasis Chakraborty
Experimental validation has been performed to show that the proposed scheme can generate new 3D structures using interpolation techniques, i. e., given two 3D structures represented as point-clouds, we can generate point-clouds in between.
no code implementations • 29 Oct 2019 • Liu Yang, Rudrasis Chakraborty, Stella X. Yu
Our proposed model is rotationally invariant and can preserve geometric shape of a 3D point-cloud.
no code implementations • 29 Oct 2019 • Liu Yang, Sean Treichler, Thorsten Kurth, Keno Fischer, David Barajas-Solano, Josh Romero, Valentin Churavy, Alexandre Tartakovsky, Michael Houston, Prabhat, George Karniadakis
Uncertainty quantification for forward and inverse problems is a central challenge across physical and biomedical disciplines.
no code implementations • 29 Aug 2019 • Liu Yang, George Em. Karniadakis
We propose a potential flow generator with $L_2$ optimal transport regularity, which can be easily integrated into a wide range of generative models including different versions of GANs and flow-based models.
2 code implementations • 26 Aug 2019 • Chen Qu, Liu Yang, Minghui Qiu, Yongfeng Zhang, Cen Chen, W. Bruce Croft, Mohit Iyyer
First, we propose a positional history answer embedding method to encode conversation history with position information using BERT in a natural way.
1 code implementation • 14 May 2019 • Chen Qu, Liu Yang, Minghui Qiu, W. Bruce Croft, Yongfeng Zhang, Mohit Iyyer
One of the major challenges to multi-turn conversational search is to model the conversation history to answer the current question.
1 code implementation • 19 Apr 2019 • Liu Yang, Junjie Hu, Minghui Qiu, Chen Qu, Jianfeng Gao, W. Bruce Croft, Xiaodong Liu, Yelong Shen, Jingjing Liu
In this paper, we propose a hybrid neural conversation model that combines the merits of both response retrieval and generation methods.
no code implementations • 17 Apr 2019 • Liu Yang, Lijing Song
In this paper, we address the question answering challenge with the SQuAD 2. 0 dataset.
no code implementations • 16 Mar 2019 • Jiafeng Guo, Yixing Fan, Liang Pang, Liu Yang, Qingyao Ai, Hamed Zamani, Chen Wu, W. Bruce Croft, Xue-Qi Cheng
Ranking models lie at the heart of research on information retrieval (IR).
no code implementations • 11 Jan 2019 • Chen Qu, Liu Yang, Bruce Croft, Falk Scholer, Yongfeng Zhang
Information retrieval systems are evolving from document retrieval to answer retrieval.
1 code implementation • 11 Jan 2019 • Chen Qu, Liu Yang, Bruce Croft, Yongfeng Zhang, Johanne R. Trippas, Minghui Qiu
Due to the limited communication bandwidth in conversational search, it is important for conversational assistants to accurately detect and predict user intent in information-seeking conversations.
no code implementations • 30 Dec 2018 • Chen Qu, Feng Ji, Minghui Qiu, Liu Yang, Zhiyu Min, Haiqing Chen, Jun Huang, W. Bruce Croft
Specifically, the data selector "acts" on the source domain data to find a subset for optimization of the TL model, and the performance of the TL model can provide "rewards" in turn to update the selector.
no code implementations • 5 Nov 2018 • Liu Yang, Dongkun Zhang, George Em. Karniadakis
We developed a new class of physics-informed generative adversarial networks (PI-GANs) to solve in a unified manner forward, inverse and mixed stochastic problems based on a limited number of scattered measurements.
no code implementations • ACL 2018 • Minghui Qiu, Liu Yang, Feng Ji, Weipeng Zhao, Wei Zhou, Jun Huang, Haiqing Chen, W. Bruce Croft, Wei. Lin
Building multi-turn information-seeking conversation systems is an important and challenging research topic.
1 code implementation • 1 May 2018 • Liu Yang, Minghui Qiu, Chen Qu, Jiafeng Guo, Yongfeng Zhang, W. Bruce Croft, Jun Huang, Haiqing Chen
Our models and research findings provide new insights on how to utilize external knowledge with deep neural models for response selection and have implications for the design of the next generation of information-seeking conversation systems.
no code implementations • 23 Apr 2018 • Chen Qu, Liu Yang, W. Bruce Croft, Johanne R. Trippas, Yongfeng Zhang, Minghui Qiu
Understanding and characterizing how people interact in information-seeking conversations is crucial in developing conversational search systems.
1 code implementation • 5 Jan 2018 • Liu Yang, Qingyao Ai, Jiafeng Guo, W. Bruce Croft
As an alternative to question answering methods based on feature engineering, deep learning approaches such as convolutional neural networks (CNNs) and Long Short-Term Memory Models (LSTMs) have recently been proposed for semantic matching of questions and answers.
Ranked #12 on
Question Answering
on TrecQA
no code implementations • SIGIR 2017 • Liu Yang, Susan T. Dumais, Paul N. Bennett, Ahmed Hassan Awadallah
Email is still among the most popular online activities.
no code implementations • 17 Jul 2017 • Liu Yang, Hamed Zamani, Yongfeng Zhang, Jiafeng Guo, W. Bruce Croft
We further evaluate the neural matching models in the next question prediction task in conversations.
no code implementations • 23 Jun 2017 • Steve Hanneke, Liu Yang
We also identify the optimal dependence on the number of pieces in the query complexity of passive testing in the special case of piecewise constant functions.
no code implementations • 29 Apr 2017 • Amit Dhurandhar, Steve Hanneke, Liu Yang
In particular, we propose an approach to provably determine the time instant from which the new/changed features start becoming relevant with respect to an output variable in an agnostic (supervised) learning setting.
no code implementations • 23 Jun 2016 • Annamalai Narayanan, Liu Yang, Lihui Chen, Liu Jinliang
In order to perform scalable detection and to adapt to the drift and evolution in malware population, an online passive-aggressive classifier is used.
no code implementations • 21 Jun 2016 • Annamalai Narayanan, Guozhu Meng, Liu Yang, Jinliang Liu, Lihui Chen
To address this, we develop the Contextual Weisfeiler-Lehman kernel (CWLK) which is capable of capturing both these types of information.
no code implementations • 21 Apr 2016 • Aleksandr Y. Aravkin, Kush R. Varshney, Liu Yang
Matrix factorization is a key component of collaborative filtering-based recommendation systems because it allows us to complete sparse user-by-item ratings matrices under a low-rank assumption that encodes the belief that similar users give similar ratings and that similar items garner similar ratings.
no code implementations • 26 Dec 2015 • Steve Hanneke, Liu Yang
Under these conditions, we propose a learning method, and establish that for bounded VC subgraph classes, the cumulative excess risk grows sublinearly in the number of predictions, at a quantified rate.
no code implementations • CVPR 2015 • Liping Jing, Liu Yang, Jian Yu, Michael K. Ng
SLRM model takes advantage of the nuclear norm regularization on mapping to effectively capture the label correlations.
no code implementations • 20 May 2015 • Steve Hanneke, Varun Kanade, Liu Yang
Some of the results also describe an active learning variant of this setting, and provide bounds on the number of queries for the labels of points in the sequence sufficient to obtain the stated bounds on the error rates.
no code implementations • 20 May 2015 • Liu Yang, Steve Hanneke, Jaime Carbonell
We study the optimal rates of convergence for estimating a prior distribution over a VC class from a sequence of independent data sets respectively labeled by independent target functions sampled from the prior.
no code implementations • 3 Oct 2014 • Steve Hanneke, Liu Yang
This work establishes distribution-free upper and lower bounds on the minimax label complexity of active learning with general hypothesis classes, under various noise models.
no code implementations • NeurIPS 2013 • Liu Yang, Jaime Carbonell
We additionally study the total cost sufficient for learning, for an abstract notion of the cost of requesting the labels of a given number of examples at once.
no code implementations • 16 Jul 2012 • Steve Hanneke, Liu Yang
Specifically, it presents an active learning algorithm based on an arbitrary classification-calibrated surrogate loss function, along with an analysis of the number of label requests sufficient for the classifier returned by the algorithm to achieve a given risk under the 0-1 loss.
no code implementations • NeurIPS 2011 • Liu Yang
We study the problem of active learning in a stream-based setting, allowing the distribution of the examples to change over time.
no code implementations • NeurIPS 2008 • Liu Yang, Rong Jin, Rahul Sukthankar
For empirical evaluation, we present a direct comparison with a number of state-of-the-art methods for inductive semi-supervised learning and text categorization; and we show that SSLW results in a significant improvement in categorization accuracy, equipped with a small training set and an unlabeled resource that is weakly related to the test beds."