no code implementations • LTEDI (ACL) 2022 • Xiaotian Lin, Yingwen Fu, Ziyu Yang, Nankai Lin, Shengyi Jiang
In this paper, we report the solution of the team BERT 4EVER for the LT-EDI-2022 shared task2: Homophobia/Transphobia Detection in social media comments in ACL 2022, which aims to classify Youtube comments into one of the following categories: no, moderate, or severe depression.
no code implementations • 18 Jun 2024 • Xingming Liao, Nankai Lin, Haowen Li, Lianglun Cheng, Zhuowei Wang, Chong Chen
Compared to Flat Named Entity Recognition (FNER), annotated resources are scarce in the corpus for NNER.
no code implementations • 7 Jun 2024 • Nankai Lin, Hongyan Wu, Zhengming Chen, Zijian Li, Lianxi Wang, Shengyi Jiang, Dong Zhou, Aimin Yang
To further meet the variability (i. e., the changing of bias attributes in datasets), we reorganize datasets to follow the continuous learning setting.
no code implementations • 4 Apr 2023 • Nankai Lin, Haonan Liu, Jiajun Fang, Dong Zhou, Aimin Yang
Subsequently, our framework aligns the corresponding sentences in two legal cases to provide evidence of similarity.
no code implementations • 30 Mar 2023 • Nankai Lin, Hongbin Zhang, Menglan Shen, Yu Wang, Shengyi Jiang, Aimin Yang
Grammatical error correction (GEC) is a challenging task of natural language processing techniques.
no code implementations • 28 Mar 2023 • Nankai Lin, Junheng He, Zhenghang Tang, Dong Zhou, Aimin Yang
The multilingual text representation module uses a multilingual pre-trained language model to represent the text, the language fusion module makes the semantic spaces of different languages tend to be consistent through contrastive learning, and the text debiasing module uses contrastive learning to make the model unable to identify sensitive attributes' information.
no code implementations • 2 Feb 2023 • Xiaotian Lin, Nankai Lin, Yingwen Fu, Ziyu Yang, Shengyi Jiang
In this paper, we propose a novel self-training selection framework with two selectors to select the high-quality samples from data augmentation.
no code implementations • 1 Dec 2022 • Nankai Lin, Guanqiu Qin, Jigang Wang, Aimin Yang, Dong Zhou
We explore the effectiveness of contrastive learning for multi-label text classification tasks by the employment of these novel losses and provide a set of baseline models for deploying contrastive learning techniques on specific tasks.
no code implementations • 25 Oct 2022 • Nankai Lin, Hongyan Wu, Sihui Fu, Shengyi Jiang, Aimin Yang
Inspired by contrastive learning, we present a novel framework for Chinese spelling checking, which consists of three modules: language representation, spelling check and reverse contrastive learning.
no code implementations • 10 Sep 2022 • Nankai Lin, Sihui Fu, Hongyan Wu, Shengyi Jiang
Chinese features prominently in the Chinese communities located in the nations of Malay Archipelago.
no code implementations • 30 Apr 2022 • Nankai Lin, Xiaotian Lin, Ziyu Yang, Shengyi Jiang
In terms of the reference-based metric, we introduce sentence-level accuracy and char-level BLEU to evaluate the corrected sentences.
no code implementations • 6 Apr 2022 • Yingwen Fu, Jinyi Chen, Nankai Lin, Xixuan Huang, Xinying Qiu, Shengyi Jiang
The Yunshan Cup 2020 track focused on creating a framework for evaluating different methods of part-of-speech (POS).
no code implementations • 2 Apr 2022 • Yingwen Fu, Nankai Lin, Ziyu Yang, Shengyi Jiang
In this paper, we describe our novel dual-contrastive framework ConCNER for cross-lingual NER under the scenario of limited source-language labeled data.
1 code implementation • 2 Apr 2022 • Nankai Lin, Yingwen Fu, Xiaotian Lin, Aimin Yang, Shengyi Jiang
In the distillation XABSA task, we further explore the comparative effectiveness of different data (source dataset, translated dataset, and code-switched dataset).
Aspect-Based Sentiment Analysis Aspect-Based Sentiment Analysis (ABSA) +2
no code implementations • 3 Dec 2021 • Xiaotian Lin, Nankai Lin, Kanoksak Wattanachote, Shengyi Jiang, Lianxi Wang
On the other hand, in view of the problem that the model cannot well recognize and utilize the correlation among languages, we further proposed a language-specific representation module to enrich semantic information for the model.
no code implementations • 23 Nov 2021 • Ru Peng, Nankai Lin, Yi Fang, Shengyi Jiang, Tianyong Hao, BoYu Chen, Junbo Zhao
However, succeeding researches pointed out that limited by the uncontrolled nature of attention computation, the NMT model requires an external syntax to capture the deep syntactic awareness.
no code implementations • LREC 2022 • Nankai Lin, Yingwen Fu, Chuwei Chen, Ziyu Yang, Shengyi Jiang
In this work, we construct a text classification dataset to alleviate the resource-scare situation of the Lao language.
no code implementations • 3 Sep 2021 • Yingwen Fu, Nankai Lin, Zhihe Yang, Shengyi Jiang
In this work, we propose a dataset construction framework, which is based on labeled datasets of homologous languages and iterative optimization, to build a Malay NER dataset (MYNER) comprising 28, 991 sentences (over 384 thousand tokens).