no code implementations • EMNLP 2020 • Chengyue Jiang, Yinggong Zhao, Shanbo Chu, Libin Shen, Kewei Tu
On the other hand, symbolic rules such as regular expressions are interpretable, require no training, and often achieve decent accuracy; but rules cannot benefit from labeled data when available and hence underperform neural networks in rich-resource scenarios.
no code implementations • NAACL 2022 • Hao Zhou, Gongshen Liu, Kewei Tu
Many natural language processing tasks involve text spans and thus high-quality span representations are needed to enhance neural approaches to these tasks.
no code implementations • Findings (NAACL) 2022 • Liwen Zhang, Zixia Jia, Wenjuan Han, Zilong Zheng, Kewei Tu
Adversarial attack of structured prediction models faces various challenges such as the difficulty of perturbing discrete words, the sentence quality issue, and the sensitivity of outputs to small perturbations.
no code implementations • EMNLP 2021 • Chengyue Jiang, Zijian Jin, Kewei Tu
Neural models and symbolic rules such as regular expressions have their respective merits and weaknesses.
1 code implementation • 18 Oct 2024 • You Wu, HaoYi Wu, Kewei Tu
Recently, sharing key-value (KV) cache across layers has been found effective in efficient inference of large language models (LLMs).
no code implementations • 2 Oct 2024 • Xiang Hu, Zhihao Teng, Wei Wu, Kewei Tu
For a given input sequence, we split it into chunks and use the current chunk to retrieve past chunks for subsequent text generation.
no code implementations • 26 Jul 2024 • Chaoyi Ai, Yong Jiang, Shen Huang, Pengjun Xie, Kewei Tu
Named entity recognition (NER) models often struggle with noisy inputs, such as those with spelling mistakes or errors generated by Optical Character Recognition processes, and learning a robust NER model is challenging.
1 code implementation • 24 Jul 2024 • Yida Zhao, Chao Lou, Kewei Tu
Syntactic Transformer language models aim to achieve better generalization through simultaneously modeling syntax trees and sentences.
no code implementations • 24 Jun 2024 • Chao Lou, Zixia Jia, Zilong Zheng, Kewei Tu
Accommodating long sequences efficiently in autoregressive Transformers, especially within an extended context window, poses significant challenges due to the quadratic computational complexity and substantial KV memory requirements inherent in self-attention mechanisms.
no code implementations • 21 Jun 2024 • Qingyang Zhu, Xiang Hu, Pengyu Ji, Wei Wu, Kewei Tu
Specifically, the deep model jointly encodes internal structures and representations of words with a mechanism named $\textit{MorphOverriding}$ to ensure the indecomposability of morphemes.
1 code implementation • 17 May 2024 • HaoYi Wu, Kewei Tu
In this paper, we propose a novel method that only computes and caches the KVs of a small number of layers, thus significantly saving memory consumption and improving inference throughput.
no code implementations • 10 May 2024 • Ning Cheng, Zhaohui Yan, ZiMing Wang, Zhijie Li, Jiaming Yu, Zilong Zheng, Kewei Tu, Jinan Xu, Wenjuan Han
Large Language Models (LLMs) play a crucial role in capturing structured semantics to enhance language understanding, improve interpretability, and reduce bias.
1 code implementation • 8 Apr 2024 • Wenyang Hui, Kewei Tu
It uses a strong LLM to summarize guidelines from previous tree search experiences to enhance the ability of a weak LLM.
1 code implementation • 2 Apr 2024 • Zhuo Chen, Chengyue Jiang, Kewei Tu
In this paper, we propose a framework of utilizing interpretation methods and gold rationales to enhance models.
1 code implementation • 2 Apr 2024 • Zhuo Chen, Xinyu Wang, Yong Jiang, Pengjun Xie, Fei Huang, Kewei Tu
With our method, the origin language models can cover several times longer contexts while keeping the computing requirements close to the baseline.
2 code implementations • 13 Mar 2024 • Xiang Hu, Pengyu Ji, Qingyang Zhu, Wei Wu, Kewei Tu
A syntactic language model (SLM) incrementally generates a sentence with its syntactic tree in a left-to-right manner.
1 code implementation • 26 Nov 2023 • HaoYi Wu, Kewei Tu
Specifically, we design a conditional random field that models discrete latent representations of all words in a sentence as well as dependency arcs between them; and we use mean field variational inference for approximate inference.
1 code implementation • 9 Nov 2023 • HaoYi Wu, Wenyang Hui, Yezeng Chen, Weiqi Wu, Kewei Tu, Yi Zhou
Since the dataset only involves a narrow range of knowledge, it is easy to separately analyse the knowledge a model possesses and the reasoning ability it has.
1 code implementation • 26 Oct 2023 • Zhaohui Yan, Songlin Yang, Wei Liu, Kewei Tu
Also, most of current ERE models do not take into account higher-order interactions between multiple entities and relations, while higher-order modeling could be beneficial. In this work, we propose HyperGraph neural network for ERE ($\hgnn{}$), which is built upon the PL-marker (a state-of-the-art marker-based pipleline model).
1 code implementation • 23 Oct 2023 • Wei Liu, Songlin Yang, Yoon Kim, Kewei Tu
Scaling dense PCFGs to thousands of nonterminals via a low-rank parameterization of the rule probability tensor has been shown to be beneficial for unsupervised parsing.
1 code implementation • 18 Oct 2023 • Chao Lou, Kewei Tu
Translation-based AMR parsers have recently gained popularity due to their simplicity and effectiveness.
1 code implementation • 28 Sep 2023 • Xiang Hu, Qingyang Zhu, Kewei Tu, Wei Wu
More interestingly, the hierarchical structures induced by ReCAT exhibit strong consistency with human-annotated syntactic trees, indicating good interpretability brought by the CIO layers.
Ranked #4 on Semantic Role Labeling on OntoNotes
Constituency Grammar Induction Natural Language Inference +2
1 code implementation • 12 Sep 2023 • Weiqi Wu, Chengyue Jiang, Yong Jiang, Pengjun Xie, Kewei Tu
In this paper, we focus on probing whether PLMs store ontological knowledge and have a semantic understanding of the knowledge rather than rote memorization of the surface form.
1 code implementation • 21 Aug 2023 • Tianyu Yu, Chengyue Jiang, Chao Lou, Shen Huang, Xiaobin Wang, Wei Liu, Jiong Cai, Yangning Li, Yinghui Li, Kewei Tu, Hai-Tao Zheng, Ningyu Zhang, Pengjun Xie, Fei Huang, Yong Jiang
However, LLMs are sometimes too footloose for natural language understanding (NLU) tasks which always have restricted output and input format.
no code implementations • 1 Jul 2023 • Jiong Cai, Yong Jiang, Yue Zhang, Chengyue Jiang, Ke Yu, Jianhui Ji, Rong Xiao, Haihong Tang, Tao Wang, Zhongqiang Huang, Pengjun Xie, Fei Huang, Kewei Tu
We also show that pretraining the QE module with auto-generated QE data from user logs can further improve the overall performance.
1 code implementation • 5 Jun 2023 • Chao Lou, Kewei Tu
Neural QCFG is a grammar-based sequence-tosequence (seq2seq) model with strong inductive biases on hierarchical structures.
1 code implementation • 5 May 2023 • Zeqi Tan, Shen Huang, Zixia Jia, Jiong Cai, Yinghui Li, Weiming Lu, Yueting Zhuang, Kewei Tu, Pengjun Xie, Fei Huang, Yong Jiang
Also, we discover that the limited context length causes the retrieval knowledge to be invisible to the model.
Multilingual Named Entity Recognition named-entity-recognition +4
2 code implementations • 6 Mar 2023 • Xiang Hu, Xinyu Kong, Kewei Tu
As the structured language model learns to predict constituency trees in a self-supervised manner, only raw texts and sentence-level labels are required as training data, which makes it essentially a general constituent-level self-interpretable classification model.
1 code implementation • 8 Feb 2023 • Chengyue Jiang, Yong Jiang, Weiqi Wu, Yuting Zheng, Pengjun Xie, Kewei Tu
The subject and object noun phrases and the relation in open KG have severe redundancy and ambiguity and need to be canonicalized.
1 code implementation • 18 Dec 2022 • Chengyue Jiang, Wenyang Hui, Yong Jiang, Xiaobin Wang, Pengjun Xie, Kewei Tu
We also found MCCE is very effective in fine-grained (130 types) and coarse-grained (9 types) entity typing.
Ranked #2 on Entity Typing on Open Entity
1 code implementation • 17 Dec 2022 • Zixia Jia, Zhaohui Yan, Wenjuan Han, Zilong Zheng, Kewei Tu
Prior works on joint Information Extraction (IE) typically model instance (e. g., event triggers, entities, roles, relations) interactions by representation enhancement, type dependencies scoring, or global decoding.
1 code implementation • 3 Dec 2022 • Chengyue Jiang, Yong Jiang, Weiqi Wu, Pengjun Xie, Kewei Tu
We use mean-field variational inference for efficient type inference on very large type sets and unfold it as a neural network module to enable end-to-end training.
Ranked #3 on Entity Typing on Open Entity
1 code implementation • 3 Dec 2022 • Xinyu Wang, Jiong Cai, Yong Jiang, Pengjun Xie, Kewei Tu, Wei Lu
MoRe contains a text retrieval module and an image-based retrieval module, which retrieve related knowledge of the input text and image in the knowledge corpus respectively.
Ranked #1 on Multi-modal Named Entity Recognition on SNAP (MNER)
Multi-modal Named Entity Recognition Named Entity Recognition +3
2 code implementations • NAACL 2022 • Songlin Yang, Wei Liu, Kewei Tu
Recent research found it beneficial to use large state spaces for HMMs and PCFGs.
no code implementations • 7 Apr 2022 • Songlin Yang, Kewei Tu
Second-order semantic parsing with end-to-end mean-field inference has been shown good performance.
1 code implementation • ACL 2022 • Chao Lou, Songlin Yang, Kewei Tu
They treat nested entities as partially-observed constituency trees and propose the masked inside algorithm for partial marginalization.
1 code implementation • SemEval (NAACL) 2022 • Xinyu Wang, Yongliang Shen, Jiong Cai, Tao Wang, Xiaobin Wang, Pengjun Xie, Fei Huang, Weiming Lu, Yueting Zhuang, Kewei Tu, Wei Lu, Yong Jiang
Our system wins 10 out of 13 tracks in the MultiCoNER shared task.
Multilingual Named Entity Recognition Named Entity Recognition +1
1 code implementation • NAACL 2022 • Xinyu Wang, Min Gui, Yong Jiang, Zixia Jia, Nguyen Bach, Tao Wang, Zhongqiang Huang, Fei Huang, Kewei Tu
As text representations take the most important role in MNER, in this paper, we propose {\bf I}mage-{\bf t}ext {\bf A}lignments (ITA) to align image features into the textual space, so that the attention mechanism in transformer-based pretrained textual embeddings can be better utilized.
Ranked #1 on Multi-modal Named Entity Recognition on Twitter-17
Multi-modal Named Entity Recognition named-entity-recognition +1
1 code implementation • ACL 2022 • Songlin Yang, Kewei Tu
Constituency parsing and nested named entity recognition (NER) are similar tasks since they both aim to predict a collection of nested and non-crossing spans.
1 code implementation • Findings (ACL) 2022 • Songlin Yang, Kewei Tu
Graph-based methods, which decompose the score of a dependency tree into scores of dependency arcs, are popular in dependency parsing for decades.
1 code implementation • ACL 2022 • Songlin Yang, Kewei Tu
In a projective dependency tree, the largest subtree rooted at each word covers a contiguous sequence (i. e., a span) in the surface order.
no code implementations • ACL 2021 • Liwen Zhang, Ge Wang, Wenjuan Han, Kewei Tu
In this paper, we propose a simple yet effective method to adapt unsupervised syntactic dependency parsing methodology for unsupervised discourse dependency parsing.
no code implementations • ACL 2021 • Zechuan Hu, Yong Jiang, Nguyen Bach, Tao Wang, Zhongqiang Huang, Fei Huang, Kewei Tu
In this paper, we propose a novel unified framework for zero-shot sequence labeling with minimum risk training and design a new decomposable risk function that models the relations between the predicted labels from the source models and the true labels.
no code implementations • ACL 2021 • Zechuan Hu, Yong Jiang, Nguyen Bach, Tao Wang, Zhongqiang Huang, Fei Huang, Kewei Tu
In structured prediction problems, cross-lingual transfer learning is an efficient way to train quality models for low-resource languages, and further improvement can be obtained by learning from multiple source languages.
no code implementations • ACL (IWPT) 2021 • Xinyu Wang, Zixia Jia, Yong Jiang, Kewei Tu
This paper describes the system used in submission from SHANGHAITECH team to the IWPT 2021 Shared Task.
1 code implementation • ACL 2021 • Songlin Yang, Yanpeng Zhao, Kewei Tu
Neural lexicalized PCFGs (L-PCFGs) have been shown effective in grammar induction.
3 code implementations • ACL 2021 • Xinyu Wang, Yong Jiang, Nguyen Bach, Tao Wang, Zhongqiang Huang, Fei Huang, Kewei Tu
We find empirically that the contextual representations computed on the retrieval-based input view, constructed through the concatenation of a sentence and its external contexts, can achieve significantly improved performance compared to the original input view based only on the sentence.
Ranked #1 on Named Entity Recognition (NER) on WNUT 2017
1 code implementation • NAACL 2021 • Songlin Yang, Yanpeng Zhao, Kewei Tu
In this work, we present a new parameterization form of PCFGs based on tensor decomposition, which has at most quadratic computational complexity in the symbol number and therefore allows us to use a much larger number of symbols.
no code implementations • EACL 2021 • Kewei Tu, Yong Jiang, Wenjuan Han, Yanpeng Zhao
Unsupervised parsing learns a syntactic parser from training sentences without parse tree annotations.
no code implementations • 12 Mar 2021 • Yixian Liu, Liwen Zhang, Wenjuan Han, Yue Zhang, Kewei Tu
We focus on CommonGen, the task of generating text based on a set of concepts, as a representative task of constrained text generation.
no code implementations • COLING 2020 • Ge Wang, Kewei Tu
By relaxing the tree constraint in both the encoder and the decoder during training, we make the learning of our model fully arc-factored and thus circumvent the challenges brought by the tree constraint.
no code implementations • COLING 2020 • Ruyue Hong, Jiong Cai, Kewei Tu
Deep inside-outside recursive autoencoder (DIORA) is a neural-based model designed for unsupervised constituency parsing.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Zechuan Hu, Yong Jiang, Nguyen Bach, Tao Wang, Zhongqiang Huang, Fei Huang, Kewei Tu
The neural linear-chain CRF model is one of the most widely-used approach to sequence labeling.
no code implementations • 10 Nov 2020 • Yang Zhou, Yong Jiang, Zechuan Hu, Kewei Tu
One limitation of linear chain CRFs is their inability to model long-range dependencies between labels.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Chengyue Jiang, Zhonglin Nian, Kaihao Guo, Shanbo Chu, Yinggong Zhao, Libin Shen, Kewei Tu
Numeral embeddings represented in this manner can be plugged into existing word embedding learning approaches such as skip-gram for training.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Zhao Li, Kewei Tu
We consider the task of cross-lingual adaptation of dependency parsers without annotated target corpora and parallel corpora.
1 code implementation • COLING 2020 • Songlin Yang, Yong Jiang, Wenjuan Han, Kewei Tu
Inspired by second-order supervised dependency parsing, we proposed a second-order extension of unsupervised neural dependency models that incorporate grandparent-child or sibling information.
Ranked #1 on Dependency Grammar Induction on WSJ10
1 code implementation • ACL 2021 • Xinyu Wang, Yong Jiang, Zhaohui Yan, Zixia Jia, Nguyen Bach, Tao Wang, Zhongqiang Huang, Fei Huang, Kewei Tu
The objective function of knowledge distillation is typically the cross-entropy between the teacher and the student's output distributions.
1 code implementation • Asian Chapter of the Association for Computational Linguistics 2020 • Xinyu Wang, Kewei Tu
In this paper, we propose second-order graph-based neural dependency parsing using message passing and end-to-end neural networks.
Ranked #1 on Dependency Parsing on Chinese Treebank
2 code implementations • ACL 2021 • Xinyu Wang, Yong Jiang, Nguyen Bach, Tao Wang, Zhongqiang Huang, Fei Huang, Kewei Tu
Pretrained contextualized embeddings are powerful word representations for structured prediction tasks.
Ranked #1 on Chunking on Penn Treebank
1 code implementation • EMNLP 2020 • Wenjuan Han, Liwen Zhang, Yong Jiang, Kewei Tu
To address these problems, we propose a novel and unified framework that learns to attack a structured prediction model using a sequence-to-sequence model with feedbacks from multiple reference models of the same structured prediction task.
no code implementations • COLING 2020 • Wenjuan Han, Yong Jiang, Hwee Tou Ng, Kewei Tu
Syntactic dependency parsing is an important task in natural language processing.
1 code implementation • EMNLP 2020 • Xinyu Wang, Yong Jiang, Nguyen Bach, Tao Wang, Zhongqiang Huang, Fei Huang, Kewei Tu
The linear-chain Conditional Random Field (CRF) model is one of the most widely-used neural sequence labeling approaches.
Ranked #3 on Chunking on CoNLL 2003 (German)
no code implementations • Findings of the Association for Computational Linguistics 2020 • Xinyu Wang, Yong Jiang, Nguyen Bach, Tao Wang, Zhongqiang Huang, Fei Huang, Kewei Tu
Recent work proposes a family of contextual embeddings that significantly improves the accuracy of sequence labelers over non-contextual embeddings.
Ranked #2 on Chunking on CoNLL 2003 (German)
no code implementations • ACL 2020 • Jun Li, Yifan Cao, Jiong Cai, Yong Jiang, Kewei Tu
Unsupervised constituency parsing aims to learn a constituency parser from a training corpus without parse tree annotations.
1 code implementation • ACL 2020 • Zixia Jia, Youmi Ma, Jiong Cai, Kewei Tu
Semantic dependency parsing, which aims to find rich bi-lexical relationships, allows words to have multiple dependency heads, resulting in graph-structured representations.
1 code implementation • ACL 2020 • Bo Pang, Erik Nijkamp, Wenjuan Han, Linqi Zhou, Yixian Liu, Kewei Tu
Open-domain dialogue generation has gained increasing attention in Natural Language Processing.
1 code implementation • WS 2020 • Xinyu Wang, Yong Jiang, Kewei Tu
This paper presents the system used in our submission to the \textit{IWPT 2020 Shared Task}.
1 code implementation • CONLL 2019 • Xinyu Wang, Yixian Liu, Zixia Jia, Chengyue Jiang, Kewei Tu
This paper presents the system used in our submission to the \textit{CoNLL 2019 shared task: Cross-Framework Meaning Representation Parsing}.
1 code implementation • ACL 2020 • Xinyu Wang, Yong Jiang, Nguyen Bach, Tao Wang, Fei Huang, Kewei Tu
Multilingual sequence labeling is a task of predicting label sequences using a single unified model for multiple languages.
no code implementations • 28 Dec 2019 • Chengyue Jiang, Zhonglin Nian, Kaihao Guo, Shanbo Chu, Yinggong Zhao, Libin Shen, Kewei Tu
Numeral embeddings represented in this manner can be plugged into existing word embedding learning approaches such as skip-gram for training.
no code implementations • IJCNLP 2019 • Yong Jiang, Wenjuan Han, Kewei Tu
Grammar induction aims to discover syntactic structures from unannotated sentences.
no code implementations • IJCNLP 2019 • Wenjuan Han, Ge Wang, Yong Jiang, Kewei Tu
The key to multilingual grammar induction is to couple grammar parameters of different languages together by exploiting the similarity between languages.
1 code implementation • AAAI 2019 • Yunzhe Yuan, Yong Jiang, Kewei Tu
Traditionally, a transitionbased dependency parser processes an input sentence and predicts a sequence of parsing actions in a left-to-right manner.
no code implementations • ACL 2019 • Wenjuan Han, Yong Jiang, Kewei Tu
In this paper, we propose a novel probabilistic model called discriminative neural dependency model with valence (D-NDMV) that generates a sentence and its parse from a continuous latent representation, which encodes global contextual information of the generated sentence.
Ranked #2 on Dependency Grammar Induction on WSJ10
Constituency Grammar Induction Dependency Grammar Induction +2
1 code implementation • ACL 2019 • Liwen Zhang, Kewei Tu, Yue Zhang
Neural models have been investigated for sentiment classification over constituent trees.
4 code implementations • ACL 2019 • Xinyu Wang, Jingxian Huang, Kewei Tu
Semantic dependency parsing aims to identify semantic relationships between words in a sentence that form a graph.
Ranked #3 on Semantic Dependency Parsing on DM
no code implementations • 13 Aug 2018 • Yanpeng Zhao, Wei Bi, Deng Cai, Xiaojiang Liu, Kewei Tu, Shuming Shi
Then, by recombining the content with the target style, we decode a sentence aligned in the target domain.
1 code implementation • ACL 2018 • Yanpeng Zhao, Liwen Zhang, Kewei Tu
We introduce Latent Vector Grammars (LVeGs), a new framework that extends latent variable grammars such that each nonterminal symbol is associated with a continuous vector space representing the set of (infinitely many) subtypes of the nonterminal.
1 code implementation • 10 Apr 2018 • Lin Qiu, Hao Zhou, Yanru Qu, Wei-Nan Zhang, Suoheng Li, Shu Rong, Dongyu Ru, Lihua Qian, Kewei Tu, Yong Yu
Information Extraction (IE) refers to automatically extracting structured relation tuples from unstructured texts.
1 code implementation • EMNLP 2017 • Xiao Zhang, Yong Jiang, Hao Peng, Kewei Tu, Dan Goldwasser
In this paper we propose an end-to-end neural CRF autoencoder (NCRF-AE) model for semi-supervised learning of sequential structured prediction problems.
no code implementations • 16 Aug 2017 • Jun Mei, Yong Jiang, Kewei Tu
For the theoretical part, we reduce general MAP inference to its special case without evidence and hidden variables; we also show that it is NP-hard to approximate the MAP problem to $2^{n^\epsilon}$ for fixed $0 \leq \epsilon < 1$, where $n$ is the input size.
1 code implementation • ICCV 2017 • Chen Zhu, Yanpeng Zhao, Shuaiyi Huang, Kewei Tu, Yi Ma
In this paper, we demonstrate the importance of encoding such relations by showing the limited effective receptive field of ResNet on two datasets, and propose to model the visual attention as a multivariate distribution over a grid-structured Conditional Random Field on image regions.
1 code implementation • EMNLP 2017 • Jiong Cai, Yong Jiang, Kewei Tu
The encoder part of our model is discriminative and globally normalized which allows us to use rich features as well as universal linguistic priors.
Dependency Grammar Induction Unsupervised Dependency Parsing
no code implementations • EMNLP 2017 • Wenjuan Han, Yong Jiang, Kewei Tu
We study the impact of big models (in terms of the degree of lexicalization) and big data (in terms of the training corpus size) on dependency grammar induction.
no code implementations • EMNLP 2017 • Yong Jiang, Wenjuan Han, Kewei Tu
Unsupervised dependency parsing aims to learn a dependency parser from unannotated sentences.
no code implementations • 8 Sep 2016 • Shanbo Chu, Yong Jiang, Kewei Tu
Probabilistic modeling is one of the foundations of modern machine learning and artificial intelligence.
no code implementations • 2 Jun 2015 • Kewei Tu
Stochastic And-Or grammars (AOG) extend traditional stochastic grammars of language to model other types of data such as images and events.
no code implementations • 2 Oct 2014 • Maria Pavlovskaia, Kewei Tu, Song-Chun Zhu
In many statistical learning problems, the target functions to be optimized are highly non-convex in various model spaces and thus are difficult to analyze.
no code implementations • NeurIPS 2013 • Kewei Tu, Maria Pavlovskaia, Song-Chun Zhu
Stochastic And-Or grammars compactly represent both compositionality and reconfigurability and have been used to model different types of data such as images and events.
no code implementations • 29 Aug 2013 • Kewei Tu, Meng Meng, Mun Wai Lee, Tae Eun Choe, Song-Chun Zhu
We present a probabilistic generative model for joint parsing that captures the relations between the input video/text, their corresponding parse graphs and the joint parse graph.