no code implementations • 10 Apr 2024 • Yunlong Feng, Yang Xu, Libo Qin, Yasheng Wang, Wanxiang Che
The framework motivates the model itself to automatically generate rationales on existing datasets.
1 code implementation • 17 Mar 2024 • Honglin Mu, Yang Xu, Yunlong Feng, Xiaofeng Han, Yitong Li, Yutai Hou, Wanxiang Che
With the rise of Large Language Models (LLMs), AI assistants' ability to utilize tools, especially through API calls, has advanced notably.
1 code implementation • 26 Jan 2024 • Haochen Tan, Zhijiang Guo, Zhan Shi, Lu Xu, Zhili Liu, Yunlong Feng, Xiaoguang Li, Yasheng Wang, Lifeng Shang, Qun Liu, Linqi Song
LLMs are prompted to generate extensive content in response to these meta-questions.
1 code implementation • 17 May 2023 • Libo Qin, Qiguang Chen, Xiao Xu, Yunlong Feng, Wanxiang Che
Spoken Language Understanding (SLU) is one of the core components of a task-oriented dialogue system, which aims to extract the semantic meaning of user queries (e. g., intents and slots).
no code implementations • 19 Apr 2023 • Bohan Li, Longxu Dou, Yutai Hou, Yunlong Feng, Honglin Mu, Qingfu Zhu, Qinghua Sun, Wanxiang Che
Prompt-based learning has shown considerable promise in reformulating various downstream tasks as cloze problems by combining original input with a predetermined template.
no code implementations • 18 Apr 2023 • Yunlong Feng, Bohan Li, Libo Qin, Xiao Xu, Wanxiang Che
Cross-domain text classification aims to adapt models to a target domain that lacks labeled data.
no code implementations • 4 Feb 2023 • Bohan Li, Xiao Xu, Xinghao Wang, Yutai Hou, Yunlong Feng, Feng Wang, Xuanliang Zhang, Qingfu Zhu, Wanxiang Che
In contrast, generative methods bring more image diversity in the augmented images but may not preserve semantic consistency, thus incorrectly changing the essential semantics of the original image.
no code implementations • CONLL 2020 • Longxu Dou, Yunlong Feng, Yuqiu Ji, Wanxiang Che, Ting Liu
This paper describes our submission system (HIT-SCIR) for the CoNLL 2020 shared task: Cross-Framework and Cross-Lingual Meaning Representation Parsing.
no code implementations • 29 Sep 2020 • Yunlong Feng, Qiang Wu
Furthermore, it is shown that several well-known robust nonconvex regression paradigms, such as Tukey regression and truncated least square regression, can be reformulated into this new framework.
no code implementations • 27 Sep 2020 • Yunlong Feng, Qiang Wu
Third, with an adaptive choice of the scale parameter, we demonstrate that Huber regression estimators can be asymptotic mean regression calibrated under $(1+\epsilon)$-moment conditions ($\epsilon>0$).
1 code implementation • EMNLP (ACL) 2021 • Wanxiang Che, Yunlong Feng, Libo Qin, Ting Liu
We introduce \texttt{N-LTP}, an open-source neural language technology platform supporting six fundamental Chinese NLP tasks: {lexical analysis} (Chinese word segmentation, part-of-speech tagging, and named entity recognition), {syntactic parsing} (dependency parsing), and {semantic parsing} (semantic dependency parsing and semantic role labeling).
no code implementations • 19 Jun 2020 • Yunlong Feng
Third, we present some new results when it is utilized to learn the conditional mean function by developing its error bounds and exponential convergence rates under conditional $(1+\epsilon)$-moment assumptions.
1 code implementation • 3 May 2020 • Yuning Yang, Yunlong Feng
In this paper, based on the maximum a posterior estimation, we derive a robust orthogonal tensor CPD model with Cauchy loss, which is resistant to heavy-tailed noise or outliers.
Optimization and Control
no code implementations • 1 Mar 2018 • Yunlong Feng, Yiming Ying
Motivated by the practical way of generating non-Gaussian noise or outliers, we introduce mixture of symmetric stable noise, which include Gaussian noise, Cauchy noise, and their mixture as special cases, to model non-Gaussian noise or outliers.
no code implementations • 20 Feb 2017 • Yunlong Feng, Jun Fan, Johan A. K. Suykens
However, it outperforms these regression models in terms of robustness as shown in our study from a re-descending M-estimation view.
no code implementations • 13 Jul 2016 • Hanyuan Hang, Ingo Steinwart, Yunlong Feng, Johan A. K. Suykens
We study the density estimation problem with observations generated by certain dynamical systems that admit a unique underlying invariant Lebesgue density.
no code implementations • 10 May 2016 • Hanyuan Hang, Yunlong Feng, Ingo Steinwart, Johan A. K. Suykens
We show that when the stochastic processes satisfy a generalized Bernstein-type inequality, a unified treatment on analyzing the learning schemes with various mixing processes can be conducted and a sharp oracle inequality for generic regularized empirical risk minimization schemes can be established.