no code implementations • 5 Feb 2024 • Benjamin Colburn, Luis G. Sanchez Giraldo, Kan Li, Jose C. Principe
We provide an extended functional Wiener equation, and present a solution to this equation in an explicit, finite dimensional, data-dependent RKHS.
1 code implementation • 19 Jan 2024 • Yiwei Li, Peiwen Yuan, Shaoxiong Feng, Boyuan Pan, Xinglin Wang, Bin Sun, HeDa Wang, Kan Li
Self-consistency (SC) has been a widely used decoding strategy for chain-of-thought reasoning.
1 code implementation • 19 Jan 2024 • Peiwen Yuan, Xinglin Wang, Shaoxiong Feng, Boyuan Pan, Yiwei Li, HeDa Wang, Xupeng Miao, Kan Li
Memorizing-free matching mechanism from Dense Retrieval (DR) is then introduced to conduct fine-grained intra-cluster matching from clusters to relevant documents.
1 code implementation • 31 Dec 2023 • Peiwen Yuan, Shaoxiong Feng, Yiwei Li, Xinglin Wang, Boyuan Pan, HeDa Wang, Kan Li
Significant progress has been made in automatic text evaluation with the introduction of large language models (LLMs) as evaluators.
1 code implementation • 20 Dec 2023 • Yiwei Li, Peiwen Yuan, Shaoxiong Feng, Boyuan Pan, Bin Sun, Xinglin Wang, HeDa Wang, Kan Li
In this work, we illustrate the merit of negative data and propose a model specialization framework to distill LLMs with negative samples besides positive ones.
1 code implementation • 17 Oct 2023 • Hang Yin, Pinren Lu, Ziang Li, Bin Sun, Kan Li
The need for high-quality data has been a key issue hindering the research of dialogue tasks.
no code implementations • 5 Sep 2023 • Jiawen Liu, Kan Li
To enhance sampling diversity and improve the model's adaptability, we propose a smooth function that maps the combined result of sentence-level and word-level information to an appropriate range, and employ probabilistic sampling based on the mapped values instead of threshold truncation.
no code implementations • 23 May 2023 • Mingzhu Zhang, Ruiping Yin, Zhen Yang, Yipeng Wang, Kan Li
Multi-task learning has been widely applied in computational vision, natural language processing and other fields, which has achieved well performance.
no code implementations • 21 Mar 2023 • Yiwei Li, Shaoxiong Feng, Bin Sun, Kan Li
Collaborative learning, also known as online knowledge distillation, is an effective way to conduct one-stage group distillation in the absence of a well-trained large teacher model.
no code implementations • 2 Dec 2022 • Bin Sun, Yitong Li, Fei Mi, Weichao Wang, Yiwei Li, Kan Li
Specifically, HLV constrains the global semantics of responses through discrete latent variables and enriches responses with continuous latent variables.
no code implementations • 1 Dec 2022 • Bin Sun, Shaoxiong Feng, Yiwei Li, Weichao Wang, Fei Mi, Yitong Li, Kan Li
Complex dialogue mappings (CDM), including one-to-many and many-to-one mappings, tend to make dialogue models generate incoherent or dull responses, and modeling these mappings remains a huge challenge for neural dialogue systems.
no code implementations • 23 May 2022 • Yiwei Li, Bin Sun, Shaoxiong Feng, Kan Li
However, the discarded samples may obtain high scores in other perspectives and can provide regularization effects on the model learning, which causes the performance improvement to be sensitive to the filtering ratio.
no code implementations • NAACL 2022 • Yiwei Li, Shaoxiong Feng, Bin Sun, Kan Li
Generative dialogue models suffer badly from the generic response problem, limiting their applications to a few toy scenarios.
no code implementations • Findings (ACL) 2022 • Shaoxiong Feng, Xuancheng Ren, Kan Li, Xu sun
However, for the continual increase of online chit-chat scenarios, directly fine-tuning these models for each of the new tasks not only explodes the capacity of the dialogue system on the embedded devices but also causes knowledge forgetting on pre-trained models and knowledge interference among diverse dialogue tasks.
no code implementations • ACL 2021 • Bin Sun, Shaoxiong Feng, Yiwei Li, Jiamou Liu, Kan Li
Conditional Variational AutoEncoder (CVAE) effectively increases the diversity and informativeness of responses in open-ended dialogue generation tasks through enriching the context vector with sampled latent variables.
no code implementations • 28 May 2021 • Bin Sun, Shaoxiong Feng, Yiwei Li, Jiamou Liu, Kan Li
In this work, we proposed a conversation model named "THINK" (Teamwork generation Hover around Impressive Noticeable Keywords) to make the decoder more complicated and avoid generating duplicated and self-contradicting responses.
no code implementations • 22 Feb 2021 • Shaoxiong Feng, Xuancheng Ren, Kan Li, Xu sun
The finding of general knowledge is further hindered by the unidirectional distillation, as the student should obey the teacher and may discard some knowledge that is truly general but refuted by the teacher.
no code implementations • EMNLP 2020 • Shaoxiong Feng, Xuancheng Ren, Hongshen Chen, Bin Sun, Kan Li, Xu sun
Human dialogues are scenario-based and appropriate responses generally relate to the latent context knowledge entailed by the specific scenario.
no code implementations • 16 Sep 2020 • Shaoxiong Feng, Hongshen Chen, Xuancheng Ren, Zhuoye Ding, Kan Li, Xu sun
Collaborative learning has successfully applied knowledge transfer to guide a pool of small student networks towards robust local minima.
no code implementations • 4 Mar 2020 • Shaoxiong Feng, Hongshen Chen, Kan Li, Dawei Yin
Neural conversational models learn to generate responses by taking into account the dialog history.
no code implementations • 1 Jan 2020 • Kan Li, Jose C. Principe
The inner product defined by the feature mapping corresponds to a positive-definite finite-rank kernel that induces a finite-dimensional reproducing kernel Hilbert space (RKHS).
no code implementations • 10 Dec 2019 • Kan Li, Jose C. Principe
Without loss of generality, we apply this approach to classical adaptive filtering algorithms and validate the methodology to show that deterministic features are faster to generate and outperform state-of-the-art kernel methods based on random Fourier features.
no code implementations • 24 Nov 2019 • Kan Li, Jose C. Principe
We present a general nonlinear Bayesian filter for high-dimensional state estimation using the theory of reproducing kernel Hilbert space (RKHS).
no code implementations • 4 Aug 2019 • Shaojie Qu, Kan Li, Zheyi Fan, Sisi Wu, Xinyi Liu, Zhiguo Huang
This paper mainly focuses on the students' accomplishment behaviors in programming assignments and compiled information from programming assignments.
no code implementations • 15 Jul 2018 • Shuai Jiang, Kan Li, Richard Yi Da Xu
Low rank matrix factorisation is often used in recommender systems as a way of extracting latent features.
1 code implementation • 5 Mar 2018 • Shuai Jiang, Kan Li, Richard Yida Xu
Non-negative Matrix Factorisation (NMF) has been extensively used in machine learning and data analytics applications.