1 code implementation • ACL 2022 • Zheng Gong, Kun Zhou, Xin Zhao, Jing Sha, Shijin Wang, Ji-Rong Wen
In this paper, we study how to continually pre-train language models for improving the understanding of math problems.
no code implementations • 14 Apr 2025 • Lili Zhao, Qi Liu, Wei Chen, Liyi Chen, Ruijun Sun, Min Hou, Yang Wang, Shijin Wang
Then, we further design self-improvement strategy in target model to reduce the reliance on multiple shortcuts.
no code implementations • 20 Feb 2025 • Jiayin Lan, Jiaqi Li, Baoxin Wang, Ming Liu, Dayong Wu, Shijin Wang, Bing Qin
Using a few-shot knowledge graph construction method based on LLM, we develop NLP-AKG, an academic knowledge graph for the NLP domain, by extracting 620, 353 entities and 2, 271, 584 relations from 60, 826 papers in ACL Anthology.
no code implementations • 17 Feb 2025 • Runxuan Liu, Bei Luo, Jiaqi Li, Baoxin Wang, Ming Liu, Dayong Wu, Shijin Wang, Bing Qin
Large language models (LLMs) have shown remarkable capabilities in natural language processing.
1 code implementation • 13 Feb 2025 • Zichong Chen, Shijin Wang, Yang Zhou
Synthesizing visually impressive images that seamlessly align both text prompts and specific artistic styles remains a significant challenge in Text-to-Image (T2I) diffusion models.
no code implementations • 13 Feb 2025 • Xinxin You, Xien Liu, Qixin Sun, huan zhang, Kaiyin Zhou, Shaohui Liu, Guoping Hu, Shijin Wang, Si Liu, Ji Wu
Recent methodologies utilizing synthetic datasets have aimed to address inconsistent hallucinations in large language models (LLMs); however, these approaches are primarily tailored to specific tasks, limiting their generalizability.
no code implementations • 21 Nov 2024 • Xinjie Sun, Qi Liu, Kai Zhang, Shuanghong Shen, Fei Wang, Yan Zhuang, Zheng Zhang, Weiyin Gong, Shijin Wang, Lina Yang, Xingying Huo
To address this, we propose the Hierarchy Constraint-Aware Cognitive Diagnosis Framework (HCD), designed to more accurately represent student ability performance within real educational contexts.
1 code implementation • 9 Oct 2024 • Hao Jiang, Qi Liu, Rui Li, Shengyu Ye, Shijin Wang
In this work, we propose a new conversational framework that comprehensively integrates these information sources, collect data to train our models and evaluate their performance.
no code implementations • 23 Sep 2024 • Bin Hong, Jinze Wu, Jiayu Liu, Liang Ding, Jing Sha, Kai Zhang, Shijin Wang, Zhenya Huang
In recent years, the breakthrough of Large Language Models (LLMs) offers new ideas for achieving universal methods on graph data.
1 code implementation • 21 Sep 2024 • Yuqing Huang, Rongyang Zhang, Xuesong He, Xuyang Zhi, Hao Wang, Xin Li, Feiyang Xu, Deguang Liu, Huadong Liang, Yi Li, Jian Cui, Zimu Liu, Shijin Wang, Guoping Hu, Guiquan Liu, Qi Liu, Defu Lian, Enhong Chen
To this end, we propose \textbf{\textit{ChemEval}}, which provides a comprehensive assessment of the capabilities of LLMs across a wide range of chemical domain tasks.
1 code implementation • Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining 2024 • Haotian Zhang, Shuanghong Shen, Bihan Xu, Zhenya Huang∗, Jinze Wu, Jing Sha, Shijin Wang
Learning path recommendation aims to provide learners with a reasonable order of items to achieve their learning goals.
no code implementations • 13 Aug 2024 • Dayong Wu, Jiaqi Li, Baoxin Wang, Honghong Zhao, Siyuan Xue, Yanjie Yang, Zhijun Chang, Rui Zhang, Li Qian, Bo wang, Shijin Wang, Zhixiong Zhang, Guoping Hu
Large language models (LLMs) have shown remarkable achievements across various language tasks. To enhance the performance of LLMs in scientific literature services, we developed the scientific literature LLM (SciLit-LLM) through pre-training and supervised fine-tuning on scientific literature, building upon the iFLYTEK Spark LLM.
no code implementations • 7 Jul 2024 • Fei Wang, Weibo Gao, Qi Liu, Jiatong Li, Guanhao Zhao, Zheng Zhang, Zhenya Huang, Mengxiao Zhu, Shijin Wang, Wei Tong, Enhong Chen
Cognitive diagnosis has been developed for decades as an effective measurement tool to evaluate human cognitive status such as ability level and knowledge mastery.
1 code implementation • 23 May 2024 • Kun Zhou, Beichen Zhang, Jiapeng Wang, Zhipeng Chen, Wayne Xin Zhao, Jing Sha, Zhichao Sheng, Shijin Wang, Ji-Rong Wen
We leverage it to synthesize 6 million math problems for pre-training our JiuZhang3. 0 model, which only needs to invoke GPT-4 API 9. 3k times and pre-train on 4. 6B data.
1 code implementation • 10 May 2024 • Tong Xiao, Jiayu Liu, Zhenya Huang, Jinze Wu, Jing Sha, Shijin Wang, Enhong Chen
Knowledge System controls an implicit reasoning process, which is responsible for providing diagram information and geometry knowledge according to a step-wise reasoning goal generated by Inference System.
1 code implementation • 31 Mar 2024 • Qi Liu, Yan Zhuang, Haoyang Bi, Zhenya Huang, Weizhe Huang, Jiatong Li, Junhao Yu, Zirui Liu, Zirui Hu, Yuting Hong, Zachary A. Pardos, Haiping Ma, Mengxiao Zhu, Shijin Wang, Enhong Chen
Computerized Adaptive Testing (CAT) provides an efficient and tailored method for assessing the proficiency of examinees, by dynamically adjusting test questions based on their performance.
1 code implementation • 10 Mar 2024 • Liyang He, Zhenya Huang, Jiayu Liu, Enhong Chen, Fei Wang, Jing Sha, Shijin Wang
In this paper, we propose an innovative Bit-mask Robust Contrastive knowledge Distillation (BRCD) method, specifically devised for the distillation of semantic hashing models.
no code implementations • 9 Mar 2024 • Fei Wang, Qi Liu, Enhong Chen, Chuanren Liu, Zhenya Huang, Jinze Wu, Shijin Wang
Specifically, based on the idea of estimating the posterior distributions of cognitive diagnosis model parameters, we first provide a unified objective function for mini-batch based optimization that can be more efficiently applied to a wide range of models and large datasets.
no code implementations • 2 Nov 2023 • Keyu Ding, Yongcan Wang, Zihang Xu, Zhenzhen Jia, Shijin Wang, Cong Liu, Enhong Chen
The results demonstrate that we have achieved state-of-the-art performance for the first time in the Full-mode Key-sequence to Characters(FK2C) task.
1 code implementation • Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2023 • Wenjie Zheng, Jianfei Yu, Rui Xia, Shijin Wang
With the extracted face sequences, we propose a multimodal facial expression-aware emotion recognition model, which leverages the frame-level facial emotion distributions to help improve utterance-level emotion recognition based on multi-task learning.
Ranked #14 on
Emotion Recognition in Conversation
on MELD
Emotion Recognition in Conversation
Facial Expression Recognition (FER)
+2
1 code implementation • 27 Jun 2023 • Zihang Xu, Ziqing Yang, Yiming Cui, Shijin Wang
IDOL achieves state-of-the-art performance on ReClor and LogiQA, the two most representative benchmarks in logical reasoning MRC, and is proven to be capable of generalizing to different pre-trained models and other types of MRC benchmarks like RACE and SQuAD 2. 0 while keeping competitive general language understanding ability through testing on tasks in GLUE.
Ranked #1 on
Reading Comprehension
on ReClor
no code implementations • 19 Jun 2023 • Wayne Xin Zhao, Kun Zhou, Beichen Zhang, Zheng Gong, Zhipeng Chen, Yuanhang Zhou, Ji-Rong Wen, Jing Sha, Shijin Wang, Cong Liu, Guoping Hu
Specially, we construct a Mixture-of-Experts~(MoE) architecture for modeling mathematical text, so as to capture the common mathematical knowledge across tasks.
1 code implementation • 18 Jun 2023 • Yan Zhuang, Qi Liu, Yuting Ning, Weizhe Huang, Zachary A. Pardos, Patrick C. Kyllonen, Jiyun Zu, Qingyang Mao, Rui Lv, Zhenya Huang, Guanhao Zhao, Zheng Zhang, Shijin Wang, Enhong Chen
As AI systems continue to grow, particularly generative models like Large Language Models (LLMs), their rigorous evaluation is crucial for development and deployment.
1 code implementation • NeurIPS 2023 • Beichen Zhang, Kun Zhou, Xilin Wei, Wayne Xin Zhao, Jing Sha, Shijin Wang, Ji-Rong Wen
Based on this finding, we propose a new approach that can deliberate the reasoning steps with tool interfaces, namely \textbf{DELI}.
no code implementations • 9 May 2023 • Bo Sun, Baoxin Wang, YiXuan Wang, Wanxiang Che, Dayong Wu, Shijin Wang, Ting Liu
Our experiments show that powerful pre-trained models perform poorly on this corpus.
1 code implementation • 3 Apr 2023 • Xin Yao, Ziqing Yang, Yiming Cui, Shijin Wang
In natural language processing, pre-trained language models have become essential infrastructures.
1 code implementation • 18 Jan 2023 • Yuting Ning, Zhenya Huang, Xin Lin, Enhong Chen, Shiwei Tong, Zheng Gong, Shijin Wang
To this end, in this paper, we propose a novel contrastive pre-training approach for mathematical question representations, namely QuesCo, which attempts to bring questions with more similar purposes closer.
1 code implementation • 15 Dec 2022 • Ziqing Yang, Yiming Cui, Xin Yao, Shijin Wang
In this work, we propose a structured pruning method GRAIN (Gradient-based Intra-attention pruning), which performs task-specific pruning with knowledge distillation and yields highly effective models.
1 code implementation • 10 Nov 2022 • Yiming Cui, Wanxiang Che, Shijin Wang, Ting Liu
We propose LERT, a pre-trained language model that is trained on three types of linguistic features along with the original MLM pre-training task, using a linguistically-informed pre-training (LIP) strategy.
Ranked #6 on
Stock Market Prediction
on Astock
1 code implementation • 11 Aug 2022 • Honghong Zhao, Baoxin Wang, Dayong Wu, Wanxiang Che, Zhigang Chen, Shijin Wang
In this paper, we present an overview of the CTC 2021, a Chinese text correction task for native speakers.
1 code implementation • 13 Jun 2022 • Wayne Xin Zhao, Kun Zhou, Zheng Gong, Beichen Zhang, Yuanhang Zhou, Jing Sha, Zhigang Chen, Shijin Wang, Cong Liu, Ji-Rong Wen
Considering the complex nature of mathematical texts, we design a novel curriculum pre-training approach for improving the learning of mathematical PLMs, consisting of both basic and advanced courses.
no code implementations • 28 Feb 2022 • Ziqing Yang, Yiming Cui, Zhigang Chen, Shijin Wang
In this paper, we aim to improve the multilingual model's supervised and zero-shot performance simultaneously only with the resources from supervised languages.
no code implementations • 10 Feb 2022 • Baoxin Wang, Qingye Meng, Ziyue Wang, Honghong Zhao, Dayong Wu, Wanxiang Che, Shijin Wang, Zhigang Chen, Cong Liu
Knowledge graph embedding (KGE) models learn the representation of entities and relations in knowledge graphs.
Ranked #4 on
Link Property Prediction
on ogbl-wikikg2
1 code implementation • 26 Aug 2021 • Yiming Cui, Wei-Nan Zhang, Wanxiang Che, Ting Liu, Zhigang Chen, Shijin Wang
Achieving human-level performance on some of the Machine Reading Comprehension (MRC) datasets is no longer challenging with the help of powerful Pre-trained Language Models (PLMs).
no code implementations • ACL 2021 • Jiefu Gong, Xiao Hu, Wei Song, Ruiji Fu, Zhichao Sheng, Bo Zhu, Shijin Wang, Ting Liu
IFlyEA provides multi-level and multi-dimension analytical modules for essay assessment.
no code implementations • Joint Conference on Lexical and Computational Semantics 2021 • Ziqing Yang, Yiming Cui, Chenglei Si, Wanxiang Che, Ting Liu, Shijin Wang, Guoping Hu
Adversarial training (AT) as a regularization method has proved its effectiveness on various tasks.
1 code implementation • EMNLP (MRQA) 2021 • Ziqing Yang, Wentao Ma, Yiming Cui, Jiani Ye, Wanxiang Che, Shijin Wang
Multilingual pre-trained models have achieved remarkable performance on cross-lingual transfer learning.
1 code implementation • 10 May 2021 • Yiming Cui, Ting Liu, Wanxiang Che, Zhigang Chen, Shijin Wang
Achieving human-level performance on some of Machine Reading Comprehension (MRC) datasets is no longer challenging with the help of powerful Pre-trained Language Models (PLMs).
Ranked #1 on
Multi-Choice MRC
on ExpMRC - RACE+ (test)
no code implementations • 7 Feb 2021 • Nan Shao, Yiming Cui, Ting Liu, Shijin Wang, Guoping Hu
To deal with this challenge, most of the existing works consider paragraphs as nodes in a graph and propose graph-based methods to retrieve them.
no code implementations • 15 Jan 2021 • Haoyang Bi, Haiping Ma, Zhenya Huang, Yu Yin, Qi Liu, Enhong Chen, Yu Su, Shijin Wang
In this paper, we study a novel model-agnostic CAT problem, where we aim to propose a flexible framework that can adapt to different cognitive models.
no code implementations • COLING 2020 • Zhiqiang Guo, Zhaoci Liu, ZhenHua Ling, Shijin Wang, Lingjing Jin, Yunxia Li
Finally, a best detection accuracy of 81. 6{\%} is obtained by our proposed methods on the Mandarin AD corpus.
no code implementations • 13 Nov 2020 • Yiming Cui, Ting Liu, Shijin Wang, Guoping Hu
With the blooming of various Pre-trained Language Models (PLMs), Machine Reading Comprehension (MRC) has embraced significant improvements on various benchmarks and even surpass human performances.
1 code implementation • COLING 2020 • Wentao Ma, Yiming Cui, Chenglei Si, Ting Liu, Shijin Wang, Guoping Hu
Most pre-trained language models (PLMs) construct word representations at subword level with Byte-Pair Encoding (BPE) or its variations, by which OOV (out-of-vocab) words are almost avoidable.
no code implementations • 1 Oct 2020 • Shaolei Wang, Baoxin Wang, Jiefu Gong, Zhongyuan Wang, Xiao Hu, Xingyi Duan, Zizhuo Shen, Gang Yue, Ruiji Fu, Dayong Wu, Wanxiang Che, Shijin Wang, Guoping Hu, Ting Liu
Grammatical error diagnosis is an important task in natural language processing.
6 code implementations • Findings of the Association for Computational Linguistics 2020 • Yiming Cui, Wanxiang Che, Ting Liu, Bing Qin, Shijin Wang, Guoping Hu
Bidirectional Encoder Representations from Transformers (BERT) has shown marvelous improvements across various NLP tasks, and consecutive variants have been proposed to further improve the performance of the pre-trained language models.
Ranked #13 on
Stock Market Prediction
on Astock
1 code implementation • Findings (ACL) 2021 • Chenglei Si, Ziqing Yang, Yiming Cui, Wentao Ma, Ting Liu, Shijin Wang
To fill this important gap, we construct AdvRACE (Adversarial RACE), a new model-agnostic benchmark for evaluating the robustness of MRC models under four different types of adversarial attacks, including our novel distractor extraction and generation attacks.
1 code implementation • ACL 2020 • Wentao Ma, Yiming Cui, Ting Liu, Dong Wang, Shijin Wang, Guoping Hu
Human conversations contain many types of information, e. g., knowledge, common sense, and language habits.
1 code implementation • COLING 2020 • Yiming Cui, Ting Liu, Ziqing Yang, Zhipeng Chen, Wentao Ma, Wanxiang Che, Shijin Wang, Guoping Hu
To add diversity in this area, in this paper, we propose a new task called Sentence Cloze-style Machine Reading Comprehension (SC-MRC).
no code implementations • EMNLP 2020 • Nan Shao, Yiming Cui, Ting Liu, Shijin Wang, Guoping Hu
We construct a strong baseline model to establish that, with the proper use of pre-trained models, graph structure may not be necessary for multi-hop question answering.
1 code implementation • ACL 2020 • Ziqing Yang, Yiming Cui, Zhipeng Chen, Wanxiang Che, Ting Liu, Shijin Wang, Guoping Hu
In this paper, we introduce TextBrewer, an open-source knowledge distillation toolkit designed for natural language processing.
no code implementations • 19 Dec 2019 • Xingyi Duan, Baoxin Wang, Ziyue Wang, Wentao Ma, Yiming Cui, Dayong Wu, Shijin Wang, Ting Liu, Tianxiang Huo, Zhen Hu, Heng Wang, Zhiyuan Liu
We present a Chinese judicial reading comprehension (CJRC) dataset which contains approximately 10K documents and almost 50K questions with answers.
no code implementations • 19 Dec 2019 • Yiming Cui, Wanxiang Che, Wei-Nan Zhang, Ting Liu, Shijin Wang, Guoping Hu
Story Ending Prediction is a task that needs to select an appropriate ending for the given story, which requires the machine to understand the story and sometimes needs commonsense knowledge.
no code implementations • 14 Nov 2019 • Yiming Cui, Wei-Nan Zhang, Wanxiang Che, Ting Liu, Zhipeng Chen, Shijin Wang, Guoping Hu
Recurrent Neural Networks (RNN) are known as powerful models for handling sequential data, and especially widely utilized in various natural language processing tasks.
no code implementations • 9 Nov 2019 • Ziqing Yang, Yiming Cui, Wanxiang Che, Ting Liu, Shijin Wang, Guoping Hu
With virtual adversarial training (VAT), we explore the possibility of improving the RC models with semi-supervised learning and prove that examples from a different task are also beneficial.
no code implementations • IJCNLP 2019 • Ziyue Wang, Baoxin Wang, Xingyi Duan, Dayong Wu, Shijin Wang, Guoping Hu, Ting Liu
To our knowledge, IFlyLegal is the first Chinese legal system that employs up-to-date NLP techniques and caters for needs of different user groups, such as lawyers, judges, procurators, and clients.
no code implementations • CONLL 2019 • Wentao Ma, Yiming Cui, Nan Shao, Su He, Wei-Nan Zhang, Ting Liu, Shijin Wang, Guoping Hu
The heart of TripleNet is a novel attention mechanism named triple attention to model the relationships within the triple at four levels.
1 code implementation • IJCNLP 2019 • Yiming Cui, Wanxiang Che, Ting Liu, Bing Qin, Shijin Wang, Guoping Hu
In this paper, we propose Cross-Lingual Machine Reading Comprehension (CLMRC) task for the languages other than English.
2 code implementations • 23 Aug 2019 • Fei Wang, Qi Liu, Enhong Chen, Zhenya Huang, Yuying Chen, Yu Yin, Zai Huang, Shijin Wang
Cognitive diagnosis is a fundamental issue in intelligent education, which aims to discover the proficiency level of students on specific knowledge concepts.
no code implementations • 27 May 2019 • Yu Yin, Qi Liu, Zhenya Huang, Enhong Chen, Wei Tong, Shijin Wang, Yu Su
Then we propose a two-level hierarchical pre-training algorithm to learn better understanding of test questions in an unsupervised way.
no code implementations • 23 May 2019 • Qi Liu, Shiwei Tong, Chuanren Liu, Hongke Zhao, Enhong Chen, Haiping Ma, Shijin Wang
Although it is well known that modeling the cognitive structure including knowledge level of learners and knowledge structure (e. g., the prerequisite relations) of learning items is important for learning path recommendation, existing methods for adaptive learning often separately focus on either knowledge levels of learners or knowledge structure of learning items.
no code implementations • 21 Nov 2018 • Zhipeng Chen, Yiming Cui, Wentao Ma, Shijin Wang, Guoping Hu
Machine Reading Comprehension (MRC) with multiple-choice questions requires the machine to read given passage and select the correct answer among several candidates.
1 code implementation • IJCNLP 2019 • Yiming Cui, Ting Liu, Wanxiang Che, Li Xiao, Zhipeng Chen, Wentao Ma, Shijin Wang, Guoping Hu
Machine Reading Comprehension (MRC) has become enormously popular recently and has attracted a lot of attention.
no code implementations • WS 2018 • Ruiji Fu, Zhengqi Pei, Jiefu Gong, Wei Song, Dechuan Teng, Wanxiang Che, Shijin Wang, Guoping Hu, Ting Liu
This paper describes our system at NLPTEA-2018 Task {\#}1: Chinese Grammatical Error Diagnosis.
no code implementations • 15 Mar 2018 • Zhipeng Chen, Yiming Cui, Wentao Ma, Shijin Wang, Ting Liu, Guoping Hu
This paper describes the system which got the state-of-the-art results at SemEval-2018 Task 11: Machine Comprehension using Commonsense Knowledge.
1 code implementation • LREC 2018 • Yiming Cui, Ting Liu, Zhipeng Chen, Wentao Ma, Shijin Wang, Guoping Hu
Machine Reading Comprehension (MRC) has become enormously popular recently and has attracted a lot of attention.
2 code implementations • ACL 2017 • Yiming Cui, Zhipeng Chen, Si Wei, Shijin Wang, Ting Liu, Guoping Hu
Cloze-style queries are representative problems in reading comprehension.
Ranked #3 on
Question Answering
on Children's Book Test
no code implementations • COLING 2016 • Yiming Cui, Ting Liu, Zhipeng Chen, Shijin Wang, Guoping Hu
Reading comprehension has embraced a booming in recent NLP research.
no code implementations • ACL 2017 • Ting Liu, Yiming Cui, Qingyu Yin, Wei-Nan Zhang, Shijin Wang, Guoping Hu
Most existing approaches for zero pronoun resolution are heavily relying on annotated data, which is often released by shared task organizers.
no code implementations • NAACL 2016 • Yiming Cui, Shijin Wang, Jianfeng Li
Artificial neural networks are powerful models, which have been widely applied into many aspects of machine translation, such as language modeling and translation modeling.