no code implementations • EMNLP 2020 • Shen Wang, Xiaokai Wei, Cicero Nogueira dos santos, Zhiguo Wang, Ramesh Nallapati, Andrew Arnold, Bing Xiang, Philip S. Yu
Existing knowledge graph embedding approaches concentrate on modeling symmetry/asymmetry, inversion, and composition typed relations but overlook the hierarchical nature of relations.
no code implementations • 6 Jun 2023 • Kanishka Misra, Cicero Nogueira dos santos, Siamak Shakeri
Despite readily memorizing world knowledge about entities, pre-trained language models (LMs) struggle to compose together two or more facts to perform multi-hop reasoning in question-answering tasks.
no code implementations • 10 Oct 2022 • Cicero Nogueira dos santos, Zhe Dong, Daniel Cer, John Nham, Siamak Shakeri, Jianmo Ni, Yun-Hsuan Sung
The resulting soft knowledge prompts (KPs) are task independent and work as an external memory of the LMs.
no code implementations • 25 May 2022 • Dheeraj Rajagopal, Siamak Shakeri, Cicero Nogueira dos santos, Eduard Hovy, Chung-Ching Chang
Abstractive summarization systems based on pretrained language models often generate coherent but factually inconsistent sentences.
no code implementations • Findings (ACL) 2022 • Kai Hui, Honglei Zhuang, Tao Chen, Zhen Qin, Jing Lu, Dara Bahri, Ji Ma, Jai Prakash Gupta, Cicero Nogueira dos santos, Yi Tay, Don Metzler
This results in significant inference time speedups since the decoder-only architecture only needs to learn to interpret static encoder embeddings during inference.
no code implementations • Findings (ACL) 2021 • Xiaofei Ma, Cicero Nogueira dos santos, Andrew O. Arnold
The performance of state-of-the-art neural rankers can deteriorate substantially when exposed to noisy inputs or applied to a new domain.
no code implementations • 11 May 2021 • Yang Li, Ben Athiwaratkun, Cicero Nogueira dos santos, Bing Xiang
In this work, we propose to leverage the prior information embedded in pretrained language models (LM) to improve generalization for intent classification and slot labeling tasks with limited training data.
1 code implementation • ACL 2021 • Feng Nan, Cicero Nogueira dos santos, Henghui Zhu, Patrick Ng, Kathleen McKeown, Ramesh Nallapati, Dejiao Zhang, Zhiguo Wang, Andrew O. Arnold, Bing Xiang
A commonly observed problem with the state-of-the art abstractive summarization models is that the generated summaries can be factually inconsistent with the input documents.
no code implementations • EMNLP 2021 • Dheeru Dua, Cicero Nogueira dos santos, Patrick Ng, Ben Athiwaratkun, Bing Xiang, Matt Gardner, Sameer Singh
Compositional reasoning tasks like multi-hop question answering, require making latent decisions to get the final answer, given a question.
1 code implementation • EACL 2021 • Feng Nan, Ramesh Nallapati, Zhiguo Wang, Cicero Nogueira dos santos, Henghui Zhu, Dejiao Zhang, Kathleen McKeown, Bing Xiang
A key challenge for abstractive summarization is ensuring factual consistency of the generated summary with respect to the original document.
1 code implementation • ICLR 2021 • Giovanni Paolini, Ben Athiwaratkun, Jason Krone, Jie Ma, Alessandro Achille, Rishita Anubhai, Cicero Nogueira dos santos, Bing Xiang, Stefano Soatto
We propose a new framework, Translation between Augmented Natural Languages (TANL), to solve many structured prediction language tasks including joint entity and relation extraction, nested named entity recognition, relation classification, semantic role labeling, event extraction, coreference resolution, and dialogue state tracking.
Ranked #3 on
Relation Classification
on TACRED
3 code implementations • 18 Dec 2020 • Peng Shi, Patrick Ng, Zhiguo Wang, Henghui Zhu, Alexander Hanbo Li, Jun Wang, Cicero Nogueira dos santos, Bing Xiang
Most recently, there has been significant interest in learning contextual representations for various NLP tasks, by leveraging large scale text corpora to train large neural language models with self-supervised learning objectives, such as Masked Language Model (MLM).
Ranked #4 on
Text-To-SQL
on spider
1 code implementation • ACL 2021 • Yifan Gao, Henghui Zhu, Patrick Ng, Cicero Nogueira dos santos, Zhiguo Wang, Feng Nan, Dejiao Zhang, Ramesh Nallapati, Andrew O. Arnold, Bing Xiang
When multiple plausible answers are found, the system should rewrite the question for each answer to resolve the ambiguity.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Dejiao Zhang, Ramesh Nallapati, Henghui Zhu, Feng Nan, Cicero Nogueira dos santos, Kathleen McKeown, Bing Xiang
Unsupervised domain adaptation addresses the problem of leveraging labeled data in a source domain to learn a well-performing model in a target domain where labels are unavailable.
Cross-Lingual Document Classification
Document Classification
+2
no code implementations • EMNLP 2020 • Pierre L. Dognin, Igor Melnyk, Inkit Padhi, Cicero Nogueira dos santos, Payel Das
In this work, we present a dual learning approach for unsupervised text to path and path to text transfers in Commonsense Knowledge Bases (KBs).
1 code implementation • EMNLP 2020 • Siamak Shakeri, Cicero Nogueira dos santos, Henry Zhu, Patrick Ng, Feng Nan, Zhiguo Wang, Ramesh Nallapati, Bing Xiang
Our model comprises a single transformer-based encoder-decoder network that is trained end-to-end to generate both answers and questions.
no code implementations • EMNLP 2020 • Cicero Nogueira dos santos, Xiaofei Ma, Ramesh Nallapati, Zhiheng Huang, Bing Xiang
Generative models for Information Retrieval, where ranking of documents is viewed as the task of generating a query from a document's language model, were very successful in various IR tasks in the past.
1 code implementation • 22 Sep 2020 • Davis Liang, Peng Xu, Siamak Shakeri, Cicero Nogueira dos Santos, Ramesh Nallapati, Zhiheng Huang, Bing Xiang
In some cases, our model trained on synthetic data can even outperform the same model trained on real data
no code implementations • EMNLP 2020 • Ben Athiwaratkun, Cicero Nogueira dos santos, Jason Krone, Bing Xiang
We set a new state-of-the-art for few-shot slot labeling, improving substantially upon the previous 5-shot ($75. 0\% \rightarrow 90. 9\%$) and 1-shot ($70. 4\% \rightarrow 81. 0\%$) state-of-the-art results.
no code implementations • ACL 2020 • Inkit Padhi, Pierre Dognin, Ke Bai, Cicero Nogueira dos santos, Vijil Chenthamarakshan, Youssef Mroueh, Payel Das
Generative feature matching network (GFMN) is an approach for training implicit generative models for images by performing moment matching on features from pre-trained neural networks.
no code implementations • 5 Jan 2020 • Michele Merler, Cicero Nogueira dos santos, Mauro Martino, Alfio M. Gliozzo, John R. Smith
We introduce a multi-modal discriminative and generative frame-work capable of assisting humans in producing visual content re-lated to a given theme, starting from a collection of documents(textual, visual, or both).
no code implementations • ICLR 2019 • Cicero Nogueira dos Santos, Inkit Padhi, Pierre Dognin, Youssef Mroueh
We propose a non-adversarial feature matching-based approach to train generative models.
no code implementations • ICCV 2019 • Cicero Nogueira dos Santos, Youssef Mroueh, Inkit Padhi, Pierre Dognin
Perceptual features (PFs) have been used with great success in tasks such as transfer learning, style transfer, and super-resolution.
no code implementations • ACL 2018 • Cicero Nogueira dos Santos, Igor Melnyk, Inkit Padhi
We introduce a new approach to tackle the problem of offensive language in online social media.
no code implementations • ACL 2018 • Rui Zhang, Cicero Nogueira dos santos, Michihiro Yasunaga, Bing Xiang, Dragomir Radev
Coreference resolution aims to identify in a text all mentions that refer to the same real-world entity.
no code implementations • 26 Nov 2017 • Igor Melnyk, Cicero Nogueira dos santos, Kahini Wadhawan, Inkit Padhi, Abhishek Kumar
Text attribute transfer using non-parallel data requires methods that can perform disentanglement of content and linguistic attributes.
no code implementations • 7 Jul 2017 • Cicero Nogueira dos Santos, Kahini Wadhawan, Bo-Wen Zhou
We propose discriminative adversarial networks (DAN) for semi-supervised learning and loss function learning.
52 code implementations • 9 Mar 2017 • Zhouhan Lin, Minwei Feng, Cicero Nogueira dos santos, Mo Yu, Bing Xiang, Bo-Wen Zhou, Yoshua Bengio
This paper proposes a new model for extracting an interpretable sentence embedding by introducing self-attention.
4 code implementations • CONLL 2016 • Ramesh Nallapati, Bo-Wen Zhou, Cicero Nogueira dos santos, Caglar Gulcehre, Bing Xiang
In this work, we model abstractive text summarization using Attentional Encoder-Decoder Recurrent Neural Networks, and show that they achieve state-of-the-art performance on two different corpora.
Ranked #10 on
Text Summarization
on DUC 2004 Task 1
no code implementations • WS 2015 • Cicero Nogueira dos Santos, Victor Guimarães
Most state-of-the-art named entity recognition (NER) systems rely on handcrafted features and on the output of other NLP tasks such as part-of-speech (POS) tagging and text chunking.
2 code implementations • IJCNLP 2015 • Cicero Nogueira dos Santos, Bing Xiang, Bo-Wen Zhou
Relation classification is an important semantic processing task for which state-ofthe-art systems still rely on costly handcrafted features.
Ranked #27 on
Relation Extraction
on SemEval-2010 Task 8