no code implementations • EMNLP 2020 • Zhiwei Yu, Hongyu Zang, Xiaojun Wan
One of the most challenging part of recipe generation is to deal with the complex restrictions among the input ingredients.
no code implementations • CL (ACL) 2021 • Junjie Cao, Zi Lin, Weiwei Sun, Xiaojun Wan
Abstract In this work, we present a phenomenon-oriented comparative analysis of the two dominant approaches in English Resource Semantic (ERS) parsing: classic, knowledge-intensive and neural, data-intensive models.
no code implementations • EMNLP 2020 • Zhiwei Yu, Hongyu Zang, Xiaojun Wan
Punning is a creative way to make conversation enjoyable and literary writing elegant.
no code implementations • ICLR 2019 • Ke Wang, Xiaojun Wan
In this paper, we propose a novel adversarial learning framework, namely DelibGAN, for generating high-quality sentences without supervision.
1 code implementation • NAACL 2022 • Mingqi Gao, Xiaojun Wan
Dialogue summarization is receiving increasing attention from researchers due to its extraordinary difficulty and unique application value.
1 code implementation • ACL 2022 • Xunjian Yin, Xiaojun Wan
With the rapid development of deep learning, Seq2Seq paradigm has become prevalent for end-to-end data-to-text generation, and the BLEU scores have been increasing in recent years.
no code implementations • EMNLP 2021 • Yitao Cai, Yue Cao, Xiaojun Wan
Concretely, we transform a sentence into a variety of different semantic or syntactic representations (including AMR, UD, and latent semantic representation), and then decode the sentence back from the semantic representations.
1 code implementation • 27 Oct 2023 • Yuchen Shen, Xiaojun Wan
Opinion summarization sets itself apart from other types of summarization tasks due to its distinctive focus on aspects and sentiments.
no code implementations • 25 Oct 2023 • Xiang Chen, Xiaojun Wan
Results illuminate LLMs' capacity and deficiency to incorporate constraints and provide insights for future developments in constrained text generation.
1 code implementation • 23 Oct 2023 • Xunjian Yin, Baizhou Huang, Xiaojun Wan
With the rapid development of NLP, large-scale language models (LLMs) excel in various tasks across multiple domains now.
1 code implementation • 10 Oct 2023 • Shiping Yang, Renliang Sun, Xiaojun Wan
Contrasting previous studies of zero-resource hallucination detection, our method and benchmark concentrate on passage-level detection instead of sentence-level.
1 code implementation • 8 Oct 2023 • Xiang Chen, Zheng Li, Xiaojun Wan
In this paper, we study the problem of controlled text editing by natural language instruction.
no code implementations • 29 Sep 2023 • Baizhou Huang, Shuai Lu, Weizhu Chen, Xiaojun Wan, Nan Duan
Specifically, we ask LLMs to sample multiple diverse outputs from various perspectives for a given query and then construct a multipartite graph based on them.
no code implementations • 18 Sep 2023 • Xiao Pu, Mingqi Gao, Xiaojun Wan
How well can large language models (LLMs) generate summaries?
no code implementations • 25 Jul 2023 • Xunjian Yin, Xiaojun Wan
With the development of pre-trained models and the incorporation of phonetic and graphic information, neural models have achieved high scores in Chinese Spelling Check (CSC).
no code implementations • 1 Jul 2023 • Huixuan Zhang, Xiaojun Wan
In this paper, we focus on multimodal hyperbole detection.
1 code implementation • 8 Jun 2023 • Mingqi Gao, Xiaojun Wan, Jia Su, Zhefeng Wang, Baoxing Huai
To address this problem, we are the first to manually annotate a FEC dataset for dialogue summarization containing 4000 items and propose FERRANTI, a fine-grained evaluation framework based on reference correction that automatically evaluates the performance of FEC models on different error categories.
1 code implementation • 7 Jun 2023 • Shiping Yang, Renliang Sun, Xiaojun Wan
Sentence Simplification is a valuable technique that can benefit language learners and children a lot.
no code implementations • 24 May 2023 • Xiao Pu, Mingqi Gao, Xiaojun Wan
The results show that summaries generated by fine-tuned models lead to higher consistency in usefulness across all three tasks, as rankings of fine-tuned summarization systems are close across downstream tasks according to the proposed extrinsic metrics.
1 code implementation • 21 May 2023 • Renliang Sun, Wei Xu, Xiaojun Wan
In this paper, we propose a new continued pre-training strategy to teach the pre-trained model to generate simple texts.
no code implementations • 2 May 2023 • Anya Belz, Craig Thomson, Ehud Reiter, Gavin Abercrombie, Jose M. Alonso-Moral, Mohammad Arvan, Anouck Braggaar, Mark Cieliebak, Elizabeth Clark, Kees Van Deemter, Tanvi Dinkar, Ondřej Dušek, Steffen Eger, Qixiang Fang, Mingqi Gao, Albert Gatt, Dimitra Gkatzia, Javier González-Corbelle, Dirk Hovy, Manuela Hürlimann, Takumi Ito, John D. Kelleher, Filip Klubicka, Emiel Krahmer, Huiyuan Lai, Chris van der Lee, Yiru Li, Saad Mahamood, Margot Mieskes, Emiel van Miltenburg, Pablo Mosteiro, Malvina Nissim, Natalie Parde, Ondřej Plátek, Verena Rieser, Jie Ruan, Joel Tetreault, Antonio Toral, Xiaojun Wan, Leo Wanner, Lewis Watson, Diyi Yang
We report our efforts in identifying a set of previous human evaluations in NLP that would be suitable for a coordinated study examining what makes human evaluations in NLP more/less reproducible.
1 code implementation • 5 Apr 2023 • Mingqi Gao, Jie Ruan, Renliang Sun, Xunjian Yin, Shiping Yang, Xiaojun Wan
Evaluating text summarization is a challenging problem, and existing evaluation metrics are far from satisfactory.
no code implementations • 6 Mar 2023 • Hui Liu, Xiaojun Wan
In this work, we conduct a detailed human evaluation of the factuality in video captioning and collect two annotated factuality datasets.
1 code implementation • 14 Feb 2023 • Renliang Sun, Zhixian Yang, Xiaojun Wan
One of the major problems with text simplification is the lack of high-quality data.
no code implementations • 20 Nov 2022 • Jie Ruan, Yue Wu, Xiaojun Wan, Yuesheng Zhu
Sarcasm generation has been investigated in previous studies by considering it as a text-to-text generation problem, i. e., generating a sarcastic sentence for an input sentence.
no code implementations • 15 Nov 2022 • Xunjian Yin, Xinyu Hu, Xiaojun Wan
Chinese Spelling Check (CSC) aims to detect and correct error tokens in Chinese contexts, which has a wide range of applications.
no code implementations • 17 Oct 2022 • Mingqi Gao, Xiaojun Wan
Many studies have revealed that word embeddings, language models, and models for specific downstream tasks in NLP are prone to social biases, especially gender bias.
no code implementations • 16 Sep 2022 • Xu Zhang, Xiaojun Wan
In view of the importance of data augmentation in APE, we separately study the impact of the construction method of artificial corpora and artificial data domain on the performance of APE models.
no code implementations • 3 Sep 2022 • Baizhou Huang, Shikang Du, Xiaojun Wan
Crosstalk is a traditional Chinese theatrical performance art.
2 code implementations • 28 Jun 2022 • Fan Xu, Yunxiang Zhang, Xiaojun Wan
Solving Chinese character riddles is a challenging task that demands understanding of character glyph, general knowledge, and a grasp of figurative language.
1 code implementation • NAACL 2022 • Zhixian Yang, Renliang Sun, Xiaojun Wan
k-nearest-neighbor machine translation (NN-MT), proposed by Khandelwal et al. (2021), has achieved many state-of-the-art results in machine translation tasks.
no code implementations • 16 Apr 2022 • Renliang Sun, Xiaojun Wan
We use a small-scale simple text dataset for continued pre-training and employ two methods to identify simple words from the texts.
1 code implementation • ACL 2022 • Zhixian Yang, Xiaojun Wan
Various models have been proposed to incorporate knowledge of syntactic structures into neural language models.
1 code implementation • 29 Jan 2022 • Xiang Chen, Xiaojun Wan
Aspect-based sentiment analysis (ABSA) is a fine-grained sentiment analysis task which aims to extract the aspects from sentences and identify their corresponding sentiments.
Aspect-Based Sentiment Analysis
Aspect-Based Sentiment Analysis (ABSA)
+4
1 code implementation • COLING 2022 • Zhe Lin, Xiaojun Wan
Zero-shot paraphrase generation has drawn much attention as the large-scale high-quality paraphrase corpus is limited.
no code implementations • 27 Dec 2021 • Yuan YAO, Qingxiu Dong, Jian Guan, Boxi Cao, Zhengyan Zhang, Chaojun Xiao, Xiaozhi Wang, Fanchao Qi, Junwei Bao, Jinran Nie, Zheni Zeng, Yuxian Gu, Kun Zhou, Xuancheng Huang, Wenhao Li, Shuhuai Ren, Jinliang Lu, Chengqiang Xu, Huadong Wang, Guoyang Zeng, Zile Zhou, Jiajun Zhang, Juanzi Li, Minlie Huang, Rui Yan, Xiaodong He, Xiaojun Wan, Xin Zhao, Xu sun, Yang Liu, Zhiyuan Liu, Xianpei Han, Erhong Yang, Zhifang Sui, Maosong Sun
We argue that for general-purpose language intelligence evaluation, the benchmark itself needs to be comprehensive and systematic.
no code implementations • 16 Dec 2021 • Sheng Xu, Xiaojun Wan
Then we propose a three-step framework to tackle this task and focus on the content extraction step in this study.
no code implementations • 5 Nov 2021 • Zhaohong Wan, Xiaojun Wan
However, these methods lack the use of syntactic knowledge which plays an important role in the correction of grammatical errors.
1 code implementation • EMNLP 2021 • Renliang Sun, Hanqi Jin, Xiaojun Wan
Finally, we select several representative models as baseline models for this task and perform automatic evaluation and human evaluation.
no code implementations • 23 Sep 2021 • Yunxiang Zhang, Xiaojun Wan
A riddle is a question or statement with double or veiled meanings, followed by an unexpected answer.
1 code implementation • Findings (EMNLP) 2021 • Chenxiao Liu, Xiaojun Wan
We propose CodeQA, a free-form question answering dataset for the purpose of source code comprehension: given a code snippet and a question, a textual answer is required to be generated.
1 code implementation • NAACL 2022 • Yunxiang Zhang, Xiaojun Wan
In this paper, we tackle the challenging task of hyperbole generation to transfer a literal sentence into its hyperbolic paraphrase.
1 code implementation • Findings (EMNLP) 2021 • Zhe Lin, Yitao Cai, Xiaojun Wan
Paraphrase generation is an important task in natural language processing.
1 code implementation • Findings (ACL) 2021 • Zhe Lin, Xiaojun Wan
Both automatic and human evaluation show BTmPG can improve the diversity of paraphrase while preserving the semantics of the original sentence.
no code implementations • ACL 2021 • Hui Liu, Xiaojun Wan
Most previous methods simplify this task by using ground-truth event segments.
1 code implementation • Findings (ACL) 2021 • Yitao Cai, Zhe Lin, Xiaojun Wan
We argue that the misprediction of concepts is due to the high relevance between English tokens and AMR concepts.
no code implementations • NAACL 2021 • Yue Cao, Hao-Ran Wei, Boxing Chen, Xiaojun Wan
In practical applications, NMT models are usually trained on a general domain corpus and then fine-tuned by continuing training on the in-domain corpus.
no code implementations • AAAI 2021 • Ke Wang, Guandan Chen, Zhongqiang Huang, Xiaojun Wan, Fei Huang
Despite the near-human performances already achieved on formal texts such as news articles, neural machine transla- tion still has difficulty in dealing with ”user-generated” texts that have diverse linguistic phenomena but lack large-scale high-quality parallel corpora.
1 code implementation • COLING 2022 • Zhixian Yang, Pengxuan Xu, Xiaojun Wan
Neural text generation models are likely to suffer from the low-diversity problem.
no code implementations • 14 Feb 2021 • Shaowei Yao, Jiwei Tan, Xi Chen, Keping Yang, Rong Xiao, Hongbo Deng, Xiaojun Wan
We propose a novel way to consider samples of different relevance confidence, and come up with a new training objective to learn a robust relevance model with desirable score distribution.
1 code implementation • EACL 2021 • Qingxiu Dong, Xiaojun Wan, Yue Cao
We propose ParaSCI, the first large-scale paraphrase dataset in the scientific field, including 33, 981 paraphrase pairs from ACL (ParaSCI-ACL) and 316, 063 pairs from arXiv (ParaSCI-arXiv).
1 code implementation • COLING 2020 • Renliang Sun, Zhe Lin, Xiaojun Wan
Our model uses neural networks to learn the different effects of the preceding sentences and the following sentences on the current sentence and applies them to the improved transformer model.
no code implementations • COLING 2020 • Zhaohong Wan, Xiaojun Wan, Wenguang Wang
The incorporation of data augmentation method in grammatical error correction task has attracted much attention.
2 code implementations • EMNLP 2020 • Yitao Cai, Xiaojun Wan
Our model outperforms previous state-of-the-art model by a large margin and achieves new state-of-the-art results on the two datasets.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Hanqi Jin, Xiaojun Wan
Single-document and multi-document summarizations are very closely related in both task definition and solution method.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Yue Cao, Xiaojun Wan
In this paper, we propose a deep generative model to generate diverse paraphrases.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Ke Wang, Xiaojun Wan
In this paper, we propose a sequence contrast loss driven text generation framework, which learns the difference between real texts and generated texts and uses that difference.
no code implementations • 7 Sep 2020 • Zilong Wang, Zhaohong Wan, Xiaojun Wan
Enlightened by recent success of Transformer in the area of machine translation, we propose a new fusion method, TransModality, to address the task of multimodal sentiment analysis.
Ranked #1 on
Multimodal Sentiment Analysis
on CMU-MOSI
(F1-score (Weighted) metric)
no code implementations • 17 Jul 2020 • Yuanyuan Zhao, Weiwei Sun, Xiaojun Wan
It is reasonable to hypothesize that the divergence patterns formulated by historical linguists and typologists reflect constraints on human languages, and are thus consistent with Second Language Acquisition (SLA) in a certain way.
no code implementations • ACL 2020 • Yue Cao, Hui Liu, Xiaojun Wan
However, it is a big challenge for the model to directly learn cross-lingual summarization as it requires learning to understand different languages and learning how to summarize at the same time.
no code implementations • ACL 2020 • Zi Chai, Xiaojun Wan
Traditional Question Generation (TQG) aims to generate a question given an input passage and an answer.
no code implementations • ACL 2020 • Xinyu Xing, Xiaosheng Fan, Xiaojun Wan
In this paper, we study the challenging problem of automatic generation of citation texts in scholarly papers.
1 code implementation • ACL 2020 • Shaowei Yao, Xiaojun Wan
Multimodal Machine Translation (MMT) aims to introduce information from other modality, generally static images, to improve the translation quality.
Ranked #6 on
Multimodal Machine Translation
on Multi30K
no code implementations • ACL 2020 • Yuanyuan Zhao, Weiwei Sun, Junjie Cao, Xiaojun Wan
This paper is concerned with semantic parsing for English as a second language (ESL).
no code implementations • ACL 2020 • Shaowei Yao, Tianming Wang, Xiaojun Wan
The graph-to-sequence (Graph2Seq) learning aims to transduce graph-structured representations to word sequences for text generation.
no code implementations • ACL 2020 • Hanqi Jin, Tianming Wang, Xiaojun Wan
In this paper, we propose a multi-granularity interaction network for extractive and abstractive multi-document summarization, which jointly learn semantic representations for words, sentences, and documents.
no code implementations • TACL 2020 • Tianming Wang, Xiaojun Wan, Hanqi Jin
Abstract meaning representation (AMR)-to-text generation is the challenging task of generating natural language texts from AMR graphs, where nodes represent concepts and edges denote relations.
no code implementations • CONLL 2019 • Yanlin Feng, Xiaojun Wan
Cross-lingual sentiment analysis (CLSA) aims to improve the performance on these languages by leveraging annotated data from other languages.
2 code implementations • ACL 2019 • Hongyu Zang, Zhiwei Yu, Xiaojun Wan
In this paper, we explore a new approach for automated chess commentary generation, which aims to generate chess commentary texts in different categories (e. g., description, comparison, planning, etc.).
1 code implementation • 13 Sep 2019 • Mengdi Zhu, Zhiwei Yu, Xiaojun Wan
Ironies can not only express stronger emotions but also show a sense of humor.
no code implementations • NAACL 2019 • Hui Liu, Wentao Qin, Xiaojun Wan
So it is of vital importance to automatically synthesize a batch of news articles related to the event or topic into a new synthesis article (or overview article) for reader's convenience.
no code implementations • 4 Jul 2019 • Junjie Cao, Zi Lin, Weiwei Sun, Xiaojun Wan
We present a phenomenon-oriented comparative analysis of the two dominant approaches in task-independent semantic parsing: classic, knowledge-intensive and neural, data-intensive models.
1 code implementation • International Joint Conference on Artificial Intelligence 2019 • Tianming Wang, Xiaojun Wan
Our model uses shared attention layers for encoder and decoder, which make the most of the contextual clues, and a latent variable for learning the distribution of coherent story plots.
no code implementations • ACL 2019 • Yitao Cai, Huiyu Cai, Xiaojun Wan
We create a multi-modal sarcasm detection dataset based on Twitter.
1 code implementation • ACL 2019 • Zi Chai, Xinyu Xing, Xiaojun Wan, Bo Huang
For openQG task, we construct OQGenD, the first dataset as far as we know, and propose a model based on conditional generative adversarial networks and our question evaluation model.
1 code implementation • 3 Jun 2019 • Hongyu Zang, Xiaojun Wan
In this paper, we propose a multi-agent style transfer system (MAST) for addressing multiple style transfer tasks with limited labeled data, by leveraging abundant unlabeled data and the mutual benefit among the multiple styles.
1 code implementation • 3 Jun 2019 • Hongyu Zang, Xiaojun Wan
The low-resource (of labeled data) problem is quite common in different task generation tasks, but unlabeled data are usually abundant.
no code implementations • NAACL 2019 • Yanlin Feng, Xiaojun Wan
Our method only requires a sentiment corpus in the source language and pretrained monolingual word embeddings of both languages.
no code implementations • NAACL 2019 • Zhiwei Yu, Xiaojun Wan
In order to create novel metaphors, we propose a neural approach to metaphor generation and explore the shared inferential structure of a metaphorical usage and a literal usage of a verb.
2 code implementations • NeurIPS 2019 • Ke Wang, Hang Hua, Xiaojun Wan
Unsupervised text attribute transfer automatically transforms a text to alter a specific attribute (e. g. sentiment) without using any parallel data, while simultaneously preserving its attribute-independent content.
no code implementations • 10 Apr 2019 • Shanshan Huang, Xiaojun Wan, Xuewei Tang
Finding new academic Methods for research problems is the key task in a researcher's research career.
no code implementations • CL 2019 • Weiwei Sun, Yufei Chen, Xiaojun Wan, Meichun Liu
In this work, we propose to represent grammatical information using general directed dependency graphs.
no code implementations • WS 2018 • Jianmin Zhang, Jiwei Tan, Xiaojun Wan
In this paper, we investigate neural abstractive methods for MDS by adapting a state-of-the-art neural abstractive summarization model for SDS.
1 code implementation • CONLL 2018 • Yufei Chen, Sheng Huang, Fang Wang, Junjie Cao, Weiwei Sun, Xiaojun Wan
We present experiments for cross-domain semantic dependency analysis with a neural Maximum Subgraph parser.
1 code implementation • EMNLP 2018 • Zi Lin, Yuguang Duan, Yuan-Yuan Zhao, Weiwei Sun, Xiaojun Wan
This paper studies semantic parsing for interlanguage (L2), taking semantic role labeling (SRL) as a case task and learner Chinese as a case language.
no code implementations • COLING 2018 • Liunian Li, Xiaojun Wan
Our approach first adopts an encoder-decoder model to generate a template text with data slots to be filled and then leverages a proposed delayed copy mechanism to fill in the slots with proper data records.
no code implementations • ACL 2018 • Yajie Ye, Weiwei Sun, Xiaojun Wan
This remarkable result demonstrates the feasibility of applying a DAG transducer to resolve NLG, as well as the effectiveness of our design.
no code implementations • ACL 2018 • Yufei Chen, Weiwei Sun, Xiaojun Wan
We demonstrate that an SHRG-based parser can produce semantic graphs much more accurately than previously shown, by relating synchronous production rules to the syntacto-semantic composition process.
no code implementations • ACL 2018 • Yufei Chen, Yuan-Yuan Zhao, Weiwei Sun, Xiaojun Wan
Motivated by the positive impact of empty category on syntactic parsing, we study neural models for pre- and in-parsing detection of empty category, which has not previously been investigated.
no code implementations • ACL 2018 • Zhiwei Yu, Jiwei Tan, Xiaojun Wan
Since sequence-to-sequence models provide an effective technique for text generation, it is promising to investigate these models on the pun generation task.
no code implementations • ACL 2018 • Yitao Cai, Yin Li, Xiaojun Wan
In this paper, we focus on the task of pun location, which aims to identify the pun word in a given short text.
no code implementations • 24 Apr 2018 • Jianmin Zhang, Jiwei Tan, Xiaojun Wan
In this paper, we investigate neural abstractive methods for MDS by adapting a state-of-the-art neural abstractive summarization model for SDS.
no code implementations • 1 Nov 2017 • Shikang Du, Xiaojun Wan, Yajie Ye
Crosstalk, also known by its Chinese name xiangsheng, is a traditional Chinese comedic performing art featuring jokes and funny dialogues, and one of China's most popular cultural elements.
no code implementations • IJCNLP 2017 • Liunian Li, Xiaojun Wan, Jin-Ge Yao, Siming Yan
In this work we study the challenging task of automatically constructing essays for Chinese college entrance examination where the topic is specified in advance.
no code implementations • EMNLP 2017 • Kui Xu, Xiaojun Wan
We present the evaluation results of our universal sentiment classifier in five languages, and the results are very promising even when the parallel data between English and the target languages are not used.
no code implementations • WS 2017 • Jin-Ge Yao, Jianmin Zhang, Xiaojun Wan, Jianguo Xiao
We study the task of constructing sports news report automatically from live commentary and focus on content selection.
no code implementations • EMNLP 2017 • Jianmin Zhang, Xiaojun Wan
In this paper we investigate a new task of automatically constructing an overview article from a given set of news articles about a news event.
no code implementations • WS 2017 • Hongyu Zang, Xiaojun Wan
Data-to-text generation is very essential and important in machine writing applications.
no code implementations • EMNLP 2017 • Junjie Cao, Sheng Huang, Weiwei Sun, Xiaojun Wan
We propose a new Maximum Subgraph algorithm for first-order parsing to 1-endpoint-crossing, pagenumber-2 graphs.
no code implementations • CONLL 2017 • Weiwei Sun, Yantao Du, Xiaojun Wan
This paper is concerned with building deep grammatical relation (GR) analysis using data-driven approach.
no code implementations • CONLL 2017 • Xun Zhang, Weiwei Sun, Xiaojun Wan
This paper is concerned with whether deep syntactic information can help surface parsing, with a particular focus on empty categories.
no code implementations • ACL 2017 • Junjie Cao, Sheng Huang, Weiwei Sun, Xiaojun Wan
We study the Maximum Subgraph problem in deep dependency parsing.
no code implementations • ACL 2017 • Jiwei Tan, Xiaojun Wan, Jianguo Xiao
Abstractive summarization is the ultimate goal of document summarization research, but previously it is less investigated due to the immaturity of text generation techniques.
Ranked #12 on
Text Summarization
on CNN / Daily Mail (Anonymized)
no code implementations • ACL 2017 • Weiwei Sun, Junjie Cao, Xiaojun Wan
We model a dependency graph as a book, a particular kind of topological space, for semantic dependency parsing.
no code implementations • 17 May 2017 • Wei Wei, Xiaojun Wan
For the identification of misleading headlines, we extract features based on the congruence between headlines and bodies.
no code implementations • COLING 2016 • Jianmin Zhang, Tianming Wang, Xiaojun Wan
PKUSUMSUM is a Java platform for multilingual document summarization, and it sup-ports multiple languages, integrates 10 automatic summarization methods, and tackles three typical summarization tasks.
no code implementations • 8 Jul 2015 • Yue Hu, Xiaojun Wan
Third, we apply the extraction method and the classification model to a paper dataset in the computer science field and conduct a further analysis of the future works.
no code implementations • 8 Jul 2015 • Xiaojun Wan, Ziqiang Cao, Furu Wei, Sujian Li, Ming Zhou
However, according to our quantitative analysis, none of the existing summarization models can always produce high-quality summaries for different document sets, and even a summarization model with good overall performance may produce low-quality summaries for some document sets.
no code implementations • 8 Jul 2015 • Xiaojiang Huang, Xiaojun Wan, Jianguo Xiao
Coordinate relation refers to the relation between instances of a concept and the relation between the directly hyponyms of a concept.
no code implementations • AAAI 2015 • Xinjie Zhou, Xiaojun Wan, Jianguo Xiao
Afterwards, we propose to generate deeper and hybrid features through neural networks stacked on the word vectors.
no code implementations • TACL 2013 • Weiwei Sun, Xiaojun Wan
We present a comparative study of transition-, graph- and PCFG-based models aimed at illuminating more precisely the likely contribution of CFGs in improving Chinese dependency parsing accuracy, especially by combining heterogeneous models.