1 code implementation • 31 Jul 2024 • Zhengxuan Wu, Yuhao Zhang, Peng Qi, Yumo Xu, Rujun Han, Yian Zhang, Jifan Chen, Bonan Min, Zhiheng Huang
Surprisingly, we find that less is more, as training ReSet with high-quality, yet substantially smaller data (three-fold less) yields superior results.
no code implementations • 10 Sep 2023 • Deguang Kong, Daniel Zhou, Zhiheng Huang, Steph Sigalas
Existing neural relevance models do not give enough consideration for query and item context information which diversifies the search results to adapt for personal preference.
no code implementations • 13 Feb 2023 • Danilo Ribeiro, Shen Wang, Xiaofei Ma, Henry Zhu, Rui Dong, Deguang Kong, Juliette Burger, Anjelica Ramos, William Wang, Zhiheng Huang, George Karypis, Bing Xiang, Dan Roth
We introduce STREET, a unified multi-task and multi-domain natural language reasoning and explanation benchmark.
1 code implementation • 19 Dec 2022 • Kaiser Sun, Peng Qi, Yuhao Zhang, Lan Liu, William Yang Wang, Zhiheng Huang
We show that, with consistent tokenization, the model performs better in both in-domain and out-of-domain datasets, with a notable average of +1. 7 F2 gain when a BART model is trained on SQuAD and evaluated on 8 QA datasets.
no code implementations • 17 Dec 2022 • Jifan Chen, Yuhao Zhang, Lan Liu, Rui Dong, Xinchi Chen, Patrick Ng, William Yang Wang, Zhiheng Huang
There has been great progress in unifying various table-to-text tasks using a single encoder-decoder model trained via multi-task learning (Xie et al., 2022).
1 code implementation • 12 Oct 2022 • Xiyang Hu, Xinchi Chen, Peng Qi, Deguang Kong, Kunlun Liu, William Yang Wang, Zhiheng Huang
Multilingual information retrieval (IR) is challenging since annotated training data is costly to obtain in many languages.
1 code implementation • Findings (NAACL) 2022 • Danilo Ribeiro, Shen Wang, Xiaofei Ma, Rui Dong, Xiaokai Wei, Henry Zhu, Xinchi Chen, Zhiheng Huang, Peng Xu, Andrew Arnold, Dan Roth
Our model is able to explain a given hypothesis by systematically generating a step-by-step explanation from textual premises.
no code implementations • Findings (EMNLP) 2021 • Peng Xu, Xinchi Chen, Xiaofei Ma, Zhiheng Huang, Bing Xiang
In this work, we propose to use a graph attention network on top of the available pretrained Transformers model to learn document embeddings.
no code implementations • 12 Oct 2021 • Peng Xu, Davis Liang, Zhiheng Huang, Bing Xiang
We propose a simple strategy to obtain an extractive answer span from the generative model by leveraging the decoder cross-attention patterns.
no code implementations • 27 Sep 2021 • Zhiheng Huang, Davis Liang, Peng Xu, Bing Xiang
Transformer models, which leverage architectural improvements like self-attention, perform remarkably well on Natural Language Processing (NLP) tasks.
no code implementations • EMNLP 2020 • Cicero Nogueira dos santos, Xiaofei Ma, Ramesh Nallapati, Zhiheng Huang, Bing Xiang
Generative models for Information Retrieval, where ranking of documents is viewed as the task of generating a query from a document's language model, were very successful in various IR tasks in the past.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Zhiheng Huang, Davis Liang, Peng Xu, Bing Xiang
In this paper, we first review absolute position embeddings and existing methods for relative position embeddings.
1 code implementation • 22 Sep 2020 • Davis Liang, Peng Xu, Siamak Shakeri, Cicero Nogueira dos Santos, Ramesh Nallapati, Zhiheng Huang, Bing Xiang
In some cases, our model trained on synthetic data can even outperform the same model trained on real data
no code implementations • 17 Jul 2020 • Parminder Bhatia, Lan Liu, Kristjan Arumae, Nima Pourdamghani, Suyog Deshpande, Ben Snively, Mona Mona, Colby Wise, George Price, Shyam Ramaswamy, Xiaofei Ma, Ramesh Nallapati, Zhiheng Huang, Bing Xiang, Taha Kass-Hout
Coronavirus disease (COVID-19) has been declared as a pandemic by WHO with thousands of cases being reported each day.
no code implementations • 16 Mar 2020 • Zhiheng Huang, Peng Xu, Davis Liang, Ajay Mishra, Bing Xiang
Prior to the transformer era, bidirectional Long Short-Term Memory (BLSTM) has been the dominant modeling architecture for neural machine translation and question answering.
Ranked #1 on Text Classification on GLUE MRPC
no code implementations • 8 Apr 2019 • Zhiheng Huang, Bing Xiang
In this paper, we propose a novel way of architecture search by means of weighted networks (WeNet), which consist of a number of networks, with each assigned a weight.
1 code implementation • 22 Jan 2019 • Julian Salazar, Katrin Kirchhoff, Zhiheng Huang
The success of self-attention in NLP has led to recent applications in end-to-end encoder-decoder architectures for speech recognition.
no code implementations • 17 Jul 2018 • Davis Liang, Zhiheng Huang, Zachary C. Lipton
Despite rapid advances in speech recognition, current models remain brittle to superficial perturbations to their inputs.
no code implementations • 24 Feb 2017 • Yisen Wang, Xuejiao Deng, Songbai Pu, Zhiheng Huang
Furthermore, we introduce a CTC-based system combination, which is different from the conventional frame-wise senone-based one.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +1
1 code implementation • CVPR 2016 • Jiang Wang, Yi Yang, Junhua Mao, Zhiheng Huang, Chang Huang, Wei Xu
While deep convolutional neural networks (CNNs) have shown a great success in single-label image classification, it is important to note that real world images generally contain multiple labels, which could correspond to different objects, scenes, actions and attributes in an image.
no code implementations • NeurIPS 2015 • Haoyuan Gao, Junhua Mao, Jie zhou, Zhiheng Huang, Lei Wang, Wei Xu
The quality of the generated answers of our mQA model on this dataset is evaluated by human judges through a Turing Test.
no code implementations • CVPR 2016 • Haonan Yu, Jiang Wang, Zhiheng Huang, Yi Yang, Wei Xu
The sentence generator produces one simple short sentence that describes a specific short video interval.
25 code implementations • 9 Aug 2015 • Zhiheng Huang, Wei Xu, Kai Yu
It can also use sentence level tag information thanks to a CRF layer.
Ranked #1 on Named Entity Recognition (NER) on FindVehicle
1 code implementation • NeurIPS 2015 • Haoyuan Gao, Junhua Mao, Jie zhou, Zhiheng Huang, Lei Wang, Wei Xu
The quality of the generated answers of our mQA model on this dataset is evaluated by human judges through a Turing Test.
1 code implementation • ICCV 2015 • Junhua Mao, Wei Xu, Yi Yang, Jiang Wang, Zhiheng Huang, Alan Yuille
In particular, we propose a transposed weight sharing scheme, which not only improves performance on image captioning, but also makes the model more suitable for the novel concept learning task.
2 code implementations • 20 Dec 2014 • Junhua Mao, Wei Xu, Yi Yang, Jiang Wang, Zhiheng Huang, Alan Yuille
In this paper, we present a multimodal Recurrent Neural Network (m-RNN) model for generating novel image captions.
12 code implementations • 1 Sep 2014 • Olga Russakovsky, Jia Deng, Hao Su, Jonathan Krause, Sanjeev Satheesh, Sean Ma, Zhiheng Huang, Andrej Karpathy, Aditya Khosla, Michael Bernstein, Alexander C. Berg, Li Fei-Fei
The ImageNet Large Scale Visual Recognition Challenge is a benchmark in object category classification and detection on hundreds of object categories and millions of images.