no code implementations • 27 Sep 2024 • Liangtai Sun, Danyu Luo, Da Ma, Zihan Zhao, Baocai Chen, Zhennan Shen, Su Zhu, Lu Chen, Xin Chen, Kai Yu
We further analyze the expert layers and show that the results of expert selection vary with data from different disciplines.
no code implementations • 11 Jun 2024 • Hanqi Li, Lu Chen, Da Ma, Zijian Wu, Su Zhu, Kai Yu
In this paper, inspired by the redundancy in the parameters of large language models, we propose a novel training paradigm: Evolving Subnetwork Training (EST).
no code implementations • 3 Jun 2024 • Da Ma, Lu Chen, Pengyu Wang, Hongshen Xu, Hanqi Li, Liangtai Sun, Su Zhu, Shuai Fan, Kai Yu
Large language models (LLMs) have demonstrated proficiency across various natural language processing (NLP) tasks but often require additional training, such as continual pre-training and supervised fine-tuning.
1 code implementation • 28 Feb 2024 • Hongshen Xu, Ruisheng Cao, Su Zhu, Sheng Jiang, Hanchong Zhang, Lu Chen, Kai Yu
Previous work on spoken language understanding (SLU) mainly focuses on single-intent settings, where each input utterance merely contains one user intent.
1 code implementation • 26 Jan 2024 • Zihan Zhao, Da Ma, Lu Chen, Liangtai Sun, Zihao Li, Yi Xia, Bo Chen, Hongshen Xu, Zichen Zhu, Su Zhu, Shuai Fan, Guodong Shen, Kai Yu, Xin Chen
In its utmost form, such a generalist AI chemist could be referred to as Chemical General Intelligence.
no code implementations • 12 Jan 2023 • Jieyu Li, Lu Chen, Ruisheng Cao, Su Zhu, Hongshen Xu, Zhi Chen, Hanchong Zhang, Kai Yu
Exploring the generalization of a text-to-SQL parser is essential for a system to automatically adapt the real-world databases.
no code implementations • 10 Sep 2022 • Zhi Chen, Yuncong Liu, Lu Chen, Su Zhu, Mengyue Wu, Kai Yu
The second phase is to fine-tune the pretrained model on the TOD data.
no code implementations • 25 May 2022 • Zhi Chen, Jijia Bao, Lu Chen, Yuncong Liu, Da Ma, Bei Chen, Mengyue Wu, Su Zhu, Xin Dong, Fujiang Ge, Qingliang Miao, Jian-Guang Lou, Kai Yu
In this work, we aim to build a unified dialogue foundation model (DFM) which can be used to solve massive diverse dialogue tasks.
no code implementations • SIGDIAL (ACL) 2022 • Zhi Chen, Lu Chen, Bei Chen, Libo Qin, Yuncong Liu, Su Zhu, Jian-Guang Lou, Kai Yu
With the development of pre-trained language models, remarkable success has been witnessed in dialogue understanding (DU).
no code implementations • 9 Dec 2021 • Su Zhu, Lu Chen, Ruisheng Cao, Zhi Chen, Qingliang Miao, Kai Yu
In this paper, we propose to improve prototypical networks with vector projection distance and abstract triangular Conditional Random Field (CRF) for the few-shot NLU.
1 code implementation • ACL 2021 • Ruisheng Cao, Lu Chen, Zhi Chen, Yanbin Zhao, Su Zhu, Kai Yu
This work aims to tackle the challenging heterogeneous graph encoding problem in the text-to-SQL task.
no code implementations • NAACL 2021 • Zhi Chen, Lu Chen, Yanbin Zhao, Ruisheng Cao, Zihan Xu, Su Zhu, Kai Yu
Given a database schema, Text-to-SQL aims to translate a natural language question into the corresponding SQL query.
1 code implementation • 25 Feb 2021 • Boer Lyu, Lu Chen, Su Zhu, Kai Yu
Additionally, we adopt the word lattice graph as input to maintain multi-granularity information.
no code implementations • 22 Sep 2020 • Zhi Chen, Lu Chen, Zihan Xu, Yanbin Zhao, Su Zhu, Kai Yu
In dialogue systems, a dialogue state tracker aims to accurately find a compact representation of the current dialogue status, based on the entire dialogue history.
no code implementations • 22 Sep 2020 • Zhi Chen, Lu Chen, Yanbin Zhao, Su Zhu, Kai Yu
In task-oriented multi-turn dialogue systems, dialogue state refers to a compact representation of the user goal in the context of dialogue history.
1 code implementation • 21 Sep 2020 • Su Zhu, Ruisheng Cao, Lu Chen, Kai Yu
Few-shot slot tagging becomes appealing for rapid domain transfer and adaptation, motivated by the tremendous development of conversational dialogue systems.
no code implementations • 7 Sep 2020 • Chen Liu, Su Zhu, Lu Chen, Kai Yu
The framework consists of a slot tagging model and a rule-based value error recovery module.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +3
no code implementations • ACL 2020 • Lu Chen, Yanbin Zhao, Boer Lyu, Lesheng Jin, Zhi Chen, Su Zhu, Kai Yu
Chinese short text matching usually employs word sequences rather than character sequences to get better performance.
no code implementations • ACL 2020 • Yanbin Zhao, Lu Chen, Zhi Chen, Ruisheng Cao, Su Zhu, Kai Yu
We also adopt graph attention networks with higher-order neighborhood information to encode the rich structure in AMR graphs.
1 code implementation • ACL 2020 • Ruisheng Cao, Su Zhu, Chenyu Yang, Chen Liu, Rao Ma, Yanbin Zhao, Lu Chen, Kai Yu
One daunting problem for semantic parsing is the scarcity of annotation.
1 code implementation • 24 May 2020 • Chen Liu, Su Zhu, Zijian Zhao, Ruisheng Cao, Lu Chen, Kai Yu
In this paper, a novel BERT based SLU model (WCN-BERT SLU) is proposed to encode WCNs and the dialogue context jointly.
2 code implementations • 26 Apr 2020 • Su Zhu, Ruisheng Cao, Kai Yu
The framework is composed of dual pseudo-labeling and dual learning method, which enables an NLU model to make full use of data (labeled and unlabeled) through a closed-loop of the primal and dual tasks.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Su Zhu, Jieyu Li, Lu Chen, Kai Yu
In this paper, a novel context and schema fusion network is proposed to encode the dialogue context and schema graph by using internal and external attention mechanisms.
Ranked #8 on Multi-domain Dialogue State Tracking on MULTIWOZ 2.0
Dialogue State Tracking Multi-domain Dialogue State Tracking
no code implementations • 3 Apr 2020 • Lu Chen, Boer Lv, Chi Wang, Su Zhu, Bowen Tan, Kai Yu
For multi-domain DST, the data sparsity problem is also a major obstacle due to the increased number of state candidates.
Ranked #12 on Multi-domain Dialogue State Tracking on MULTIWOZ 2.1
no code implementations • 22 Mar 2020 • Su Zhu, Zijian Zhao, Rao Ma, Kai Yu
The proposed approaches are evaluated on three datasets.
1 code implementation • IJCNLP 2019 • Zijian Zhao, Su Zhu, Kai Yu
The atomic templates produce exemplars for fine-grained constituents of semantic representations.
1 code implementation • ACL 2019 • Ruisheng Cao, Su Zhu, Chen Liu, Jieyu Li, Kai Yu
Semantic parsing converts natural language queries into structured logical forms.
1 code implementation • 9 Apr 2019 • Zijian Zhao, Su Zhu, Kai Yu
In the paper, we focus on spoken language understanding from unaligned data whose annotation is a set of act-slot-value triples.
no code implementations • WS 2018 • Su Zhu, Kai Yu
Concept definition is important in language understanding (LU) adaptation since literal definition difference can easily lead to data sparsity even if different data sets are actually semantically correlated.
no code implementations • 6 Aug 2016 • Su Zhu, Kai Yu
This paper investigates the framework of encoder-decoder with attention for sequence labelling based spoken language understanding.