no code implementations • EMNLP 2020 • Shen Wang, Xiaokai Wei, Cicero Nogueira dos santos, Zhiguo Wang, Ramesh Nallapati, Andrew Arnold, Bing Xiang, Philip S. Yu
Existing knowledge graph embedding approaches concentrate on modeling symmetry/asymmetry, inversion, and composition typed relations but overlook the hierarchical nature of relations.
no code implementations • 1 Mar 2025 • Hui Li, Zhiguo Wang, Bohui Chen, Li Sheng
Equivariant neural networks play a pivotal role in analyzing datasets with symmetry properties, particularly in complex data structures.
no code implementations • 25 Feb 2025 • Haoyang Wen, Jiang Guo, Yi Zhang, Jiarong Jiang, Zhiguo Wang
This paper investigates synthetic data generation strategies in developing generative retrieval models for domain-specific corpora, thereby addressing the scalability challenges inherent in manually annotating in-domain queries.
no code implementations • 14 Oct 2024 • Mingwen Dong, Nischal Ashok Kumar, Yiqun Hu, Anuj Chauhan, Chung-Wei Hang, Shuaichen Chang, Lin Pan, Wuwei Lan, Henghui Zhu, Jiarong Jiang, Patrick Ng, Zhiguo Wang
Previous text-to-SQL datasets and systems have primarily focused on user questions with clear intentions that can be answered.
no code implementations • 8 Oct 2024 • Chuansen Peng, Hanning Tang, Zhiguo Wang, Xiaojing Shen
To the best of our knowledge, this is the first work to propose a first-order algorithmic framework for inferring network structures from smooth signals under partial observability, offering both guaranteed linear convergence and practical effectiveness for large-scale networks.
no code implementations • 18 Sep 2024 • Hideo Kobayashi, Wuwei Lan, Peng Shi, Shuaichen Chang, Jiang Guo, Henghui Zhu, Zhiguo Wang, Patrick Ng
While significant progress has been made on the text-to-SQL task, recent solutions repeatedly encode the same database schema for every question, resulting in unnecessary high inference cost and often overlooking crucial database knowledge.
1 code implementation • 15 Aug 2024 • Dongyu Ru, Lin Qiu, Xiangkun Hu, Tianhang Zhang, Peng Shi, Shuaichen Chang, Cheng Jiayang, Cunxiang Wang, Shichao Sun, Huanyu Li, Zizhao Zhang, Binjie Wang, Jiarong Jiang, Tong He, Zhiguo Wang, PengFei Liu, Yue Zhang, Zheng Zhang
Despite Retrieval-Augmented Generation (RAG) showing promising capability in leveraging external knowledge, a comprehensive evaluation of RAG systems is still challenging due to the modular nature of RAG, evaluation of long-form responses and reliability of measurements.
no code implementations • 24 Apr 2024 • Jiaqing Yuan, Lin Pan, Chung-Wei Hang, Jiang Guo, Jiarong Jiang, Bonan Min, Patrick Ng, Zhiguo Wang
By further decoupling model known and unknown knowledge, we find the degradation is attributed to exemplars that contradict a model's known knowledge, as well as the number of such exemplars.
1 code implementation • 31 Jan 2024 • Wenyue Hua, Jiang Guo, Mingwen Dong, Henghui Zhu, Patrick Ng, Zhiguo Wang
Our analysis over the chain-of-thought generation of edited models further uncover key reasons behind the inadequacy of existing knowledge editing methods from a reasoning standpoint, involving aspects on fact-wise editing, fact recall ability, and coherence in generation.
1 code implementation • 15 Sep 2023 • Richard Cornelius Suwandi, Zhidi Lin, Feng Yin, Zhiguo Wang, Sergios Theodoridis
This paper presents a novel GP linear multiple kernel (LMK) and a generic sparsity-aware distributed learning framework to optimize the hyper-parameters.
no code implementations • 10 Aug 2023 • Alexander Hanbo Li, Mingyue Shang, Evangelia Spiliopoulou, Jie Ma, Patrick Ng, Zhiguo Wang, Bonan Min, William Wang, Kathleen McKeown, Vittorio Castelli, Dan Roth, Bing Xiang
We present a novel approach for structured data-to-text generation that addresses the limitations of existing methods that primarily focus on specific types of structured data.
1 code implementation • 28 Jun 2023 • Kangning Yin, Zhen Ding, Zhihua Dong, Dongsheng Chen, Jie Fu, Xinhui Ji, Guangqiang Yin, Zhiguo Wang
Therefore, we open source a non-IID IoT person detection (NIPD) data set, which is collected from five different cameras.
1 code implementation • 7 Jun 2023 • Yusen Zhang, Jun Wang, Zhiguo Wang, Rui Zhang
However, existing CLSP models are separately proposed and evaluated on datasets of limited tasks and applications, impeding a comprehensive and unified evaluation of CLSP on a diverse range of NLs and MRs. To this end, we present XSemPLR, a unified benchmark for cross-lingual semantic parsing featured with 22 natural languages and 8 meaning representations by examining and selecting 9 existing datasets to cover 5 tasks and 164 domains.
no code implementations • 30 May 2023 • Xingyu Fu, Sheng Zhang, Gukyeong Kwon, Pramuditha Perera, Henghui Zhu, Yuhao Zhang, Alexander Hanbo Li, William Yang Wang, Zhiguo Wang, Vittorio Castelli, Patrick Ng, Dan Roth, Bing Xiang
The open-ended Visual Question Answering (VQA) task requires AI models to jointly reason over visual and natural language inputs using world knowledge.
no code implementations • 27 May 2023 • Sijia Wang, Alexander Hanbo Li, Henry Zhu, Sheng Zhang, Chung-Wei Hang, Pramuditha Perera, Jie Ma, William Wang, Zhiguo Wang, Vittorio Castelli, Bing Xiang, Patrick Ng
Entities can be expressed in diverse formats, such as texts, images, or column names and cell values in tables.
1 code implementation • 25 May 2023 • Wuwei Lan, Zhiguo Wang, Anuj Chauhan, Henghui Zhu, Alexander Li, Jiang Guo, Sheng Zhang, Chung-Wei Hang, Joseph Lilien, Yiqun Hu, Lin Pan, Mingwen Dong, Jun Wang, Jiarong Jiang, Stephen Ash, Vittorio Castelli, Patrick Ng, Bing Xiang
A practical text-to-SQL system should generalize well on a wide variety of natural language questions, unseen database schemas, and novel SQL query structures.
2 code implementations • 21 Jan 2023 • Shuaichen Chang, Jun Wang, Mingwen Dong, Lin Pan, Henghui Zhu, Alexander Hanbo Li, Wuwei Lan, Sheng Zhang, Jiarong Jiang, Joseph Lilien, Steve Ash, William Yang Wang, Zhiguo Wang, Vittorio Castelli, Patrick Ng, Bing Xiang
Neural text-to-SQL models have achieved remarkable performance in translating natural language questions into SQL queries.
no code implementations • 17 Dec 2022 • Yiyun Zhao, Jiarong Jiang, Yiqun Hu, Wuwei Lan, Henry Zhu, Anuj Chauhan, Alexander Li, Lin Pan, Jun Wang, Chung-Wei Hang, Sheng Zhang, Marvin Dong, Joe Lilien, Patrick Ng, Zhiguo Wang, Vittorio Castelli, Bing Xiang
In this paper, we first examined the existing synthesized datasets and discovered that state-of-the-art text-to-SQL algorithms did not further improve on popular benchmarks when trained with augmented synthetic data.
no code implementations • 3 Dec 2022 • Shuai Wang, Yanqing Xu, Zhiguo Wang, Tsung-Hui Chang, Tony Q. S. Quek, Defeng Sun
In this paper, we firstly reveal the fact that the federated ADMM is essentially a client-variance-reduced algorithm.
1 code implementation • 30 Sep 2022 • Donghan Yu, Sheng Zhang, Patrick Ng, Henghui Zhu, Alexander Hanbo Li, Jun Wang, Yiqun Hu, William Wang, Zhiguo Wang, Bing Xiang
Question answering over knowledge bases (KBs) aims to answer natural language questions with factual information such as entities and relations in KBs.
no code implementations • 28 Sep 2022 • Jun Wang, Patrick Ng, Alexander Hanbo Li, Jiarong Jiang, Zhiguo Wang, Ramesh Nallapati, Bing Xiang, Sudipta Sengupta
When synthesizing a SQL query, there is no explicit semantic information of NLQ available to the parser which leads to undesirable generalization performance.
no code implementations • 10 Jun 2022 • Sheng Zhang, Patrick Ng, Zhiguo Wang, Bing Xiang
Our generative model is a unified framework to sequentially generate relational triplets under various relation extraction settings and explicitly utilizes relevant knowledge from Knowledge Graph (KG) to resolve ambiguities.
no code implementations • 29 Mar 2022 • Yupeng Chen, Zhiguo Wang, Xiaojing Shen
Network topology inference is a fundamental problem in many applications of network science, such as locating the source of fake news, brain connectivity networks detection, etc.
no code implementations • 27 Nov 2021 • Yinchen Shen, Zhiguo Wang, Ruoyu Sun, Xiaojing Shen
Then we propose a feature selection method to reduce the size of the model, based on a new metric which trades off the classification accuracy and privacy preserving.
no code implementations • 29 Oct 2021 • Zhiguo Wang, Xintong Wang, Ruoyu Sun, Tsung-Hui Chang
Similar to that encountered in federated supervised learning, class distribution of labeled/unlabeled data could be non-i. i. d.
no code implementations • 29 Sep 2021 • Yinchen Shen, Zhiguo Wang, Ruoyu Sun, Xiaojing Shen
Differential privacy (DP) is an essential technique for privacy-preserving, which works by adding random noise to the data.
1 code implementation • ACL 2021 • Alexander Hanbo Li, Patrick Ng, Peng Xu, Henghui Zhu, Zhiguo Wang, Bing Xiang
However, a large amount of world's knowledge is stored in structured databases, and need to be accessed using query languages such as SQL.
no code implementations • 17 Jun 2021 • Peng Shi, Tao Yu, Patrick Ng, Zhiguo Wang
Furthermore, we propose two value filling methods to build the bridge from the existing zero-shot semantic parsers to real-world applications, considering most of the existing parsers ignore the values filling in the synthesized SQL.
1 code implementation • ACL 2021 • Feng Nan, Cicero Nogueira dos santos, Henghui Zhu, Patrick Ng, Kathleen McKeown, Ramesh Nallapati, Dejiao Zhang, Zhiguo Wang, Andrew O. Arnold, Bing Xiang
A commonly observed problem with the state-of-the art abstractive summarization models is that the generated summaries can be factually inconsistent with the input documents.
no code implementations • EACL 2021 • Zhiguo Wang, Patrick Ng, Ramesh Nallapati, Bing Xiang
Experiments show that: (1) Our IR-based retrieval method is able to collect high-quality candidates efficiently, thus enables our method adapt to large-scale KBs easily; (2) the BERT model improves the accuracy across all three sub-tasks; and (3) benefiting from multi-task learning, the unified model obtains further improvements with only 1/3 of the original parameters.
1 code implementation • EACL 2021 • Feng Nan, Ramesh Nallapati, Zhiguo Wang, Cicero Nogueira dos santos, Henghui Zhu, Dejiao Zhang, Kathleen McKeown, Bing Xiang
A key challenge for abstractive summarization is ensuring factual consistency of the generated summary with respect to the original document.
3 code implementations • 18 Dec 2020 • Peng Shi, Patrick Ng, Zhiguo Wang, Henghui Zhu, Alexander Hanbo Li, Jun Wang, Cicero Nogueira dos santos, Bing Xiang
Most recently, there has been significant interest in learning contextual representations for various NLP tasks, by leveraging large scale text corpora to train large neural language models with self-supervised learning objectives, such as Masked Language Model (MLM).
Ranked #7 on
Semantic Parsing
on spider
1 code implementation • ACL 2021 • Yifan Gao, Henghui Zhu, Patrick Ng, Cicero Nogueira dos santos, Zhiguo Wang, Feng Nan, Dejiao Zhang, Ramesh Nallapati, Andrew O. Arnold, Bing Xiang
When multiple plausible answers are found, the system should rewrite the question for each answer to resolve the ambiguity.
no code implementations • 10 Nov 2020 • Zhiguo Wang, Jiawei Zhang, Tsung-Hui Chang, Jian Li, Zhi-Quan Luo
While many distributed optimization algorithms have been proposed for solving smooth or convex problems over the networks, few of them can handle non-convex and non-smooth problems.
1 code implementation • EMNLP 2020 • Siamak Shakeri, Cicero Nogueira dos santos, Henry Zhu, Patrick Ng, Feng Nan, Zhiguo Wang, Ramesh Nallapati, Bing Xiang
Our model comprises a single transformer-based encoder-decoder network that is trained end-to-end to generate both answers and questions.
1 code implementation • 7 Jun 2020 • Zhiguo Wang, Liusha Yang, Feng Yin, Ke Lin, Qingjiang Shi, Zhi-Quan Luo
In this paper, we find these two methods have complementary properties and larger diversity, which motivates us to propose a new semi-supervised learning method that is able to adaptively combine the strengths of Xgboost and transductive support vector machine.
no code implementations • 24 May 2020 • Zhiguo Wang, Zhongliang Yang, Yu-Jin Zhang
First, the aggregation strategy chooses one detector as master detector by experience, and sets the remaining detectors as auxiliary detectors.
1 code implementation • ACL 2020 • Alexander R. Fabbri, Patrick Ng, Zhiguo Wang, Ramesh Nallapati, Bing Xiang
Training a QA model on this data gives a relative improvement over a previous unsupervised model in F1 score on the SQuAD dataset by about 14%, and 20% when the answer is a named entity, achieving state-of-the-art performance on SQuAD for unsupervised QA.
no code implementations • 24 Feb 2020 • Ye Li, Guangqiang Yin, Chunhui Liu, Xiaoyu Yang, Zhiguo Wang
Triplet loss processes batch construction in a complicated and fussy way and converges slowly.
1 code implementation • 25 Nov 2019 • Henghui Zhu, Feng Nan, Zhiguo Wang, Ramesh Nallapati, Bing Xiang
In this work, we define the problem of conversation structure modeling as identifying the parent utterance(s) to which each utterance in the conversation responds to.
no code implementations • 19 Nov 2019 • Zhiguo Wang, Zhongliang Yang, Yu-Jin Zhang
To address these problems, we propose a promotion method: utilize the maximum of block-level GEs on the frame to detect anomaly.
1 code implementation • IJCNLP 2019 • Linfeng Song, Yue Zhang, Daniel Gildea, Mo Yu, Zhiguo Wang, Jinsong Su
Medical relation extraction discovers relations between entity mentions in text, such as research articles.
no code implementations • WS 2019 • Xiaofei Ma, Peng Xu, Zhiguo Wang, Ramesh Nallapati, Bing Xiang
The performance of deep neural models can deteriorate substantially when there is a domain shift between training and test data.
no code implementations • 17 Oct 2019 • Xiaofei Ma, Zhiguo Wang, Patrick Ng, Ramesh Nallapati, Bing Xiang
We present a systematic investigation of layer-wise BERT activations for general-purpose text representations to understand what linguistic information they capture and how transferable they are across different tasks.
no code implementations • IJCNLP 2019 • Zhiguo Wang, Patrick Ng, Xiaofei Ma, Ramesh Nallapati, Bing Xiang
To tackle this issue, we propose a multi-passage BERT model to globally normalize answer scores across all passages of the same question, and this change enables our QA model find better answers by utilizing more passages.
Ranked #3 on
Open-Domain Question Answering
on SearchQA
no code implementations • WS 2019 • Zhiguo Wang, Yue Zhang, Mo Yu, Wei zhang, Lin Pan, Linfeng Song, Kun Xu, Yousef El-Kurdi
Self-explaining text categorization requires a classifier to make a prediction along with supporting evidence.
no code implementations • NAACL 2019 • Kun Xu, Yuxuan Lai, Yansong Feng, Zhiguo Wang
However, extending KV-MemNNs to Knowledge Based Question Answering (KB-QA) is not trivia, which should properly decompose a complex question into a sequence of queries against the memory, and update the query representations to support multi-hop reasoning over the memory.
1 code implementation • ACL 2019 • Kun Xu, Li-Wei Wang, Mo Yu, Yansong Feng, Yan Song, Zhiguo Wang, Dong Yu
Previous cross-lingual knowledge graph (KG) alignment studies rely on entity embeddings derived only from monolingual KG structural information, which may fail at matching entities that have different facts in two KGs.
1 code implementation • TACL 2019 • Linfeng Song, Daniel Gildea, Yue Zhang, Zhiguo Wang, Jinsong Su
It is intuitive that semantic representations can be useful for machine translation, mainly because they can help in enforcing meaning preservation and handling data sparsity (many sentences correspond to one meaning) of machine translation models.
no code implementations • EMNLP 2018 • Linfeng Song, Yue Zhang, Zhiguo Wang, Daniel Gildea
Cross-sentence $n$-ary relation extraction detects relations among $n$ entities across multiple sentences.
1 code implementation • EMNLP 2018 • Kun Xu, Lingfei Wu, Zhiguo Wang, Yansong Feng, Vadim Sheinin
Previous work approaches the SQL-to-text generation task using vanilla Seq2Seq models, which may not fully capture the inherent graph-structured information in SQL query.
no code implementations • 6 Sep 2018 • Linfeng Song, Zhiguo Wang, Mo Yu, Yue Zhang, Radu Florian, Daniel Gildea
Multi-hop reading comprehension focuses on one type of factoid question, where a system needs to properly integrate multiple pieces of evidence to correctly answer a question.
Ranked #2 on
Question Answering
on COMPLEXQUESTIONS
2 code implementations • 28 Aug 2018 • Linfeng Song, Yue Zhang, Zhiguo Wang, Daniel Gildea
Cross-sentence $n$-ary relation extraction detects relations among $n$ entities across multiple sentences.
1 code implementation • EMNLP 2018 • Kun Xu, Lingfei Wu, Zhiguo Wang, Mo Yu, Li-Wei Chen, Vadim Sheinin
Existing neural semantic parsers mainly utilize a sequence encoder, i. e., a sequential LSTM, to extract word order features while neglecting other valuable syntactic information such as dependency graph or constituent trees.
1 code implementation • NAACL 2018 • Linfeng Song, Zhiguo Wang, Wael Hamza, Yue Zhang, Daniel Gildea
The task of natural question generation is to generate a corresponding question given the input passage (fact) and answer.
Ranked #11 on
Question Generation
on SQuAD1.1
1 code implementation • ACL 2018 • Linfeng Song, Yue Zhang, Zhiguo Wang, Daniel Gildea
The problem of AMR-to-text generation is to recover a text representing the same meaning as an input AMR graph.
Ranked #1 on
Graph-to-Sequence
on LDC2015E86:
(using extra training data)
4 code implementations • ICLR 2019 • Kun Xu, Lingfei Wu, Zhiguo Wang, Yansong Feng, Michael Witbrock, Vadim Sheinin
Our method first generates the node and graph embeddings using an improved graph-based neural network with a novel aggregation strategy to incorporate edge direction information in the node embeddings.
Ranked #1 on
SQL-to-Text
on WikiSQL
no code implementations • 1 Mar 2018 • Yang Yu, Kazi Saidul Hasan, Mo Yu, Wei zhang, Zhiguo Wang
Relation detection is a core component for Knowledge Base Question Answering (KBQA).
1 code implementation • ICLR 2018 • Shuohang Wang, Mo Yu, Jing Jiang, Wei zhang, Xiaoxiao Guo, Shiyu Chang, Zhiguo Wang, Tim Klinger, Gerald Tesauro, Murray Campbell
We propose two methods, namely, strength-based re-ranking and coverage-based re-ranking, to make use of the aggregated evidence from different passages to better determine the answer.
Ranked #1 on
Open-Domain Question Answering
on Quasar
no code implementations • 4 Sep 2017 • Linfeng Song, Zhiguo Wang, Wael Hamza
In the QG task, a question is generated from the system given the passage and the target answer, whereas in the QA task, the answer is generated given the question and the passage.
1 code implementation • 31 Aug 2017 • Shuohang Wang, Mo Yu, Xiaoxiao Guo, Zhiguo Wang, Tim Klinger, Wei zhang, Shiyu Chang, Gerald Tesauro, Bo-Wen Zhou, Jing Jiang
Second, we propose a novel method that jointly trains the Ranker along with an answer-generation Reader model, based on reinforcement learning.
Ranked #4 on
Open-Domain Question Answering
on Quasar
no code implementations • 25 Aug 2017 • Zhiguo Wang, Wael Hamza, Linfeng Song
However, it lacks the capacity of utilizing instance-level information from individual instances in the training set.
10 code implementations • 13 Feb 2017 • Zhiguo Wang, Wael Hamza, Radu Florian
Natural language sentence matching is a fundamental technology for a variety of tasks.
Ranked #17 on
Paraphrase Identification
on Quora Question Pairs
(Accuracy metric)
no code implementations • ACL 2017 • Linfeng Song, Xiaochang Peng, Yue Zhang, Zhiguo Wang, Daniel Gildea
This paper addresses the task of AMR-to-text generation by leveraging synchronous node replacement grammar.
1 code implementation • 13 Dec 2016 • Zhiguo Wang, Haitao Mi, Wael Hamza, Radu Florian
Based on this dataset, we propose a Multi-Perspective Context Matching (MPCM) model, which is an end-to-end system that directly predicts the answer beginning and ending points in a passage.
Ranked #3 on
Open-Domain Question Answering
on SQuAD1.1
no code implementations • COLING 2016 • Young-suk Lee, Zhiguo Wang
We present a dependency to constituent tree conversion technique that aims to improve constituent parsing accuracies by leveraging dependency treebanks available in a wide variety in many languages.
no code implementations • EMNLP 2016 • Linfeng Song, Yue Zhang, Xiaochang Peng, Zhiguo Wang, Daniel Gildea
The task of AMR-to-text generation is to generate grammatical text that sustains the semantic meaning for a given AMR graph.
no code implementations • EMNLP 2016 • Haitao Mi, Zhiguo Wang, Abe Ittycheriah
We simply compute the distance between the machine attentions and the "true" alignments, and minimize this cost in the training procedure.
no code implementations • SEMEVAL 2016 • Linfeng Song, Zhiguo Wang, Haitao Mi, Daniel Gildea
In the training stage, our method induces several sense centroids (embedding) for each polysemous word.
Ranked #4 on
Word Sense Induction
on SemEval 2010 WSI
no code implementations • EMNLP 2016 • Haitao Mi, Baskaran Sankaran, Zhiguo Wang, Abe Ittycheriah
In this paper, we enhance the attention-based neural machine translation (NMT) by adding explicit coverage embedding models to alleviate issues of repeating and dropping translations in NMT.
no code implementations • ACL 2016 • Haitao Mi, Zhiguo Wang, Abe Ittycheriah
Our method simply takes into account the translation options of each word or phrase in the source sentence, and picks a very small target vocabulary for each sentence based on a word-to-word translation model or a bilingual phrase library learned from a traditional machine translation model.
1 code implementation • COLING 2016 • Zhiguo Wang, Haitao Mi, Abraham Ittycheriah
Most conventional sentence similarity methods only focus on similar parts of two input sentences, and simply ignore the dissimilar parts, which usually give us some clues and semantic meanings about the sentences.
Ranked #13 on
Question Answering
on WikiQA
no code implementations • CONLL 2016 • Zhiguo Wang, Haitao Mi, Abraham Ittycheriah
In this work, we propose a semi-supervised method for short text clustering, where we represent texts as distributed vectors with neural networks, and use a small amount of labeled data to specify our intention for clustering.
no code implementations • 9 Jul 2015 • Zhiguo Wang, Abraham Ittycheriah
In this paper, we propose a novel word-alignment-based method to solve the FAQ-based question answering task.
no code implementations • 22 Jul 2013 • Jason Satel, Ross Story, Matthew D. Hilchey, Zhiguo Wang, Raymond M. Klein
When the interval between a transient ash of light (a "cue") and a second visual response signal (a "target") exceeds at least 200ms, responding is slowest in the direction indicated by the first signal.
no code implementations • TACL 2013 • Zhiguo Wang, Cheng-qing Zong
In this paper, we take dependency cohesion as a soft constraint, and integrate it into a generative model for large-scale word alignment experiments.