Search Results for author: Yingbo Zhou

Found 64 papers, 20 papers with code

Few-Shot Intent Classification by Gauging Entailment Relationship Between Utterance and Semantic Label

no code implementations EMNLP (NLP4ConvAI) 2021 Jin Qu, Kazuma Hashimoto, Wenhao Liu, Caiming Xiong, Yingbo Zhou

Compared with DNNC, our proposed method is more efficient in both training and serving since it is based upon the entailment between query utterance and labels instead of all the training examples.

Classification intent-classification +2

Personalized Federated Instruction Tuning via Neural Architecture Search

no code implementations26 Feb 2024 Pengyu Zhang, Yingbo Zhou, Ming Hu, Junxian Feng, Jiawen Weng, Mingsong Chen

Federated Instruction Tuning (FIT) has shown the ability to achieve collaborative model instruction tuning among massive data owners without sharing private data.

Neural Architecture Search

DIVKNOWQA: Assessing the Reasoning Ability of LLMs via Open-Domain Question Answering over Knowledge Base and Text

no code implementations31 Oct 2023 Wenting Zhao, Ye Liu, Tong Niu, Yao Wan, Philip S. Yu, Shafiq Joty, Yingbo Zhou, Semih Yavuz

Moreover, a significant gap in the current landscape is the absence of a realistic benchmark for evaluating the effectiveness of grounding LLMs on heterogeneous knowledge sources (e. g., knowledge base and text).

Knowledge Graphs Open-Domain Question Answering +2

WaveAttack: Asymmetric Frequency Obfuscation-based Backdoor Attacks Against Deep Neural Networks

no code implementations17 Oct 2023 Jun Xia, Zhihao Yue, Yingbo Zhou, Zhiwei Ling, Xian Wei, Mingsong Chen

Due to the popularity of Artificial Intelligence (AI) technology, numerous backdoor attacks are designed by adversaries to mislead deep neural network predictions by manipulating training samples and training processes.

Backdoor Attack SSIM

L2CEval: Evaluating Language-to-Code Generation Capabilities of Large Language Models

no code implementations29 Sep 2023 Ansong Ni, Pengcheng Yin, Yilun Zhao, Martin Riddell, Troy Feng, Rui Shen, Stephen Yin, Ye Liu, Semih Yavuz, Caiming Xiong, Shafiq Joty, Yingbo Zhou, Dragomir Radev, Arman Cohan

Recently, large language models (LLMs), especially those that are pretrained on code, have demonstrated strong capabilities in generating programs from natural language inputs in a few-shot or even zero-shot manner.

Code Generation Math +1

Investigating Answerability of LLMs for Long-Form Question Answering

no code implementations15 Sep 2023 Meghana Moorthy Bhat, Rui Meng, Ye Liu, Yingbo Zhou, Semih Yavuz

As we embark on a new era of LLMs, it becomes increasingly crucial to understand their capabilities, limitations, and differences.

Long Form Question Answering Question Generation +1

XGen-7B Technical Report

1 code implementation7 Sep 2023 Erik Nijkamp, Tian Xie, Hiroaki Hayashi, Bo Pang, Congying Xia, Chen Xing, Jesse Vig, Semih Yavuz, Philippe Laban, Ben Krause, Senthil Purushwalkam, Tong Niu, Wojciech Kryściński, Lidiya Murakhovs'ka, Prafulla Kumar Choubey, Alex Fabbri, Ye Liu, Rui Meng, Lifu Tu, Meghana Bhat, Chien-Sheng Wu, Silvio Savarese, Yingbo Zhou, Shafiq Joty, Caiming Xiong

Most open-source LLMs, on the other hand, are limited in their ability to support longer sequence lengths, which is a key requirement for many tasks that require inference over an input context.

2k 8k

EqGAN: Feature Equalization Fusion for Few-shot Image Generation

no code implementations27 Jul 2023 Yingbo Zhou, Zhihao Yue, Yutong Ye, Pengyu Zhang, Xian Wei, Mingsong Chen

Due to the absence of fine structure and texture information, existing fusion-based few-shot image generation methods suffer from unsatisfactory generation quality and diversity.

Generative Adversarial Network Image Generation

Few-shot Unified Question Answering: Tuning Models or Prompts?

no code implementations23 May 2023 Srijan Bansal, Semih Yavuz, Bo Pang, Meghana Bhat, Yingbo Zhou

Question-answering (QA) tasks often investigate specific question types, knowledge domains, or reasoning skills, leading to specialized models catering to specific categories of QA tasks.

Question Answering Transfer Learning

HPE:Answering Complex Questions over Text by Hybrid Question Parsing and Execution

no code implementations12 May 2023 Ye Liu, Semih Yavuz, Rui Meng, Dragomir Radev, Caiming Xiong, Yingbo Zhou

It comprises two central pillars: (1) We parse the question of varying complexity into an intermediate representation, named H-expression, which is composed of simple questions as the primitives and symbolic operations representing the relationships among them; (2) To execute the resulting H-expressions, we design a hybrid executor, which integrates the deterministic rules to translate the symbolic operations with a drop-in neural reader network to answer each decomposed simple question.

Knowledge Graphs Question Answering +1

CodeGen2: Lessons for Training LLMs on Programming and Natural Languages

2 code implementations3 May 2023 Erik Nijkamp, Hiroaki Hayashi, Caiming Xiong, Silvio Savarese, Yingbo Zhou

In this study, we attempt to render the training of LLMs for program synthesis more efficient by unifying four key components: (1) model architectures, (2) learning methods, (3) infill sampling, and, (4) data distributions.

Causal Language Modeling Language Modelling +2

Efficiently Aligned Cross-Lingual Transfer Learning for Conversational Tasks using Prompt-Tuning

1 code implementation3 Apr 2023 Lifu Tu, Jin Qu, Semih Yavuz, Shafiq Joty, Wenhao Liu, Caiming Xiong, Yingbo Zhou

Our results demonstrate the strong and efficient modeling ability of NLI-based classifiers and the large cross-lingual transfer improvements achieved by our aligned prompts, particularly in few-shot settings.

Cross-Lingual Transfer intent-classification +4

CyclicFL: A Cyclic Model Pre-Training Approach to Efficient Federated Learning

no code implementations28 Jan 2023 Pengyu Zhang, Yingbo Zhou, Ming Hu, Xin Fu, Xian Wei, Mingsong Chen

Based on the concept of Continual Learning (CL), we prove that CyclicFL approximates existing centralized pre-training methods in terms of classification and prediction performance.

Continual Learning Federated Learning

AugTriever: Unsupervised Dense Retrieval by Scalable Data Augmentation

no code implementations17 Dec 2022 Rui Meng, Ye Liu, Semih Yavuz, Divyansh Agarwal, Lifu Tu, Ning Yu, JianGuo Zhang, Meghana Bhat, Yingbo Zhou

Dense retrievers have made significant strides in text retrieval and open-domain question answering, even though most achievements were made possible only with large amounts of human supervision.

Data Augmentation Open-Domain Question Answering +2

Best-$k$ Search Algorithm for Neural Text Generation

no code implementations22 Nov 2022 Jiacheng Xu, Caiming Xiong, Silvio Savarese, Yingbo Zhou

We first investigate the vanilla best-first search (BFS) algorithm and then propose the Best-$k$ Search algorithm.

Question Generation Question-Generation +2

Uni-Parser: Unified Semantic Parser for Question Answering on Knowledge Base and Database

no code implementations9 Nov 2022 Ye Liu, Semih Yavuz, Rui Meng, Dragomir Radev, Caiming Xiong, Yingbo Zhou

Parsing natural language questions into executable logical forms is a useful and interpretable way to perform question answering on structured data such as knowledge bases (KB) or databases (DB).

Question Answering Semantic Parsing

Prompt-Tuning Can Be Much Better Than Fine-Tuning on Cross-lingual Understanding With Multilingual Language Models

2 code implementations22 Oct 2022 Lifu Tu, Caiming Xiong, Yingbo Zhou

Pre-trained multilingual language models show significant performance gains for zero-shot cross-lingual model transfer on a wide range of natural language understanding (NLU) tasks.

Cross-Lingual Transfer Natural Language Understanding +3

General-to-Specific Transfer Labeling for Domain Adaptable Keyphrase Generation

1 code implementation20 Aug 2022 Rui Meng, Tong Wang, Xingdi Yuan, Yingbo Zhou, Daqing He

Finally, we fine-tune the model with limited data with true labels to fully adapt it to the target domain.

Keyphrase Generation

BigIssue: A Realistic Bug Localization Benchmark

no code implementations21 Jul 2022 Paul Kassianik, Erik Nijkamp, Bo Pang, Yingbo Zhou, Caiming Xiong

As machine learning tools progress, the inevitable question arises: How can machine learning help us write better code?

BIG-bench Machine Learning Program Repair

Modeling Multi-hop Question Answering as Single Sequence Prediction

no code implementations ACL 2022 Semih Yavuz, Kazuma Hashimoto, Yingbo Zhou, Nitish Shirish Keskar, Caiming Xiong

Fusion-in-decoder (Fid) (Izacard and Grave, 2020) is a generative question answering (QA) model that leverages passage retrieval with a pre-trained transformer and pushed the state of the art on single-hop QA.

Answer Generation Generative Question Answering +3

OneAligner: Zero-shot Cross-lingual Transfer with One Rich-Resource Language Pair for Low-Resource Sentence Retrieval

no code implementations Findings (ACL) 2022 Tong Niu, Kazuma Hashimoto, Yingbo Zhou, Caiming Xiong

When finetuned on a single rich-resource language pair, be it English-centered or not, our model is able to match the performance of the ones finetuned on all language pairs under the same data budget with less than 2. 0 points decrease in accuracy.

Machine Translation Retrieval +2

CodeGen: An Open Large Language Model for Code with Multi-Turn Program Synthesis

5 code implementations25 Mar 2022 Erik Nijkamp, Bo Pang, Hiroaki Hayashi, Lifu Tu, Huan Wang, Yingbo Zhou, Silvio Savarese, Caiming Xiong

To democratize this, we train and release a family of large language models up to 16. 1B parameters, called CODEGEN, on natural language and programming language data, and open source the training library JAXFORMER.

Code Generation Language Modelling +2

Long Document Summarization with Top-down and Bottom-up Inference

no code implementations15 Mar 2022 Bo Pang, Erik Nijkamp, Wojciech Kryściński, Silvio Savarese, Yingbo Zhou, Caiming Xiong

Critical to the success of a summarization model is the faithful inference of latent representations of words or tokens in the source documents.

Choose Your QA Model Wisely: A Systematic Study of Generative and Extractive Readers for Question Answering

no code implementations SpaNLP (ACL) 2022 Man Luo, Kazuma Hashimoto, Semih Yavuz, Zhiwei Liu, Chitta Baral, Yingbo Zhou

Among several interesting findings, it is important to highlight that (1) the generative readers perform better in long context QA, (2) the extractive readers perform better in short context while also showing better out-of-domain generalization, and (3) the encoder of encoder-decoder PrLMs (e. g., T5) turns out to be a strong extractive reader and outperforms the standard choice of encoder-only PrLMs (e. g., RoBERTa).

Domain Generalization Multi-Task Learning +1

Efficient and Differentiable Conformal Prediction with General Function Classes

1 code implementation ICLR 2022 Yu Bai, Song Mei, Huan Wang, Yingbo Zhou, Caiming Xiong

Experiments show that our algorithm is able to learn valid prediction sets and improve the efficiency significantly over existing approaches in several applications such as prediction intervals with improved length, minimum-volume prediction sets for multi-output regression, and label prediction sets for image classification.

Conformal Prediction Image Classification +2

Dense Hierarchical Retrieval for Open-Domain Question Answering

1 code implementation Findings (EMNLP) 2021 Ye Liu, Kazuma Hashimoto, Yingbo Zhou, Semih Yavuz, Caiming Xiong, Philip S. Yu

In this work, we propose Dense Hierarchical Retrieval (DHR), a hierarchical framework that can generate accurate dense representations of passages by utilizing both macroscopic semantics in the document and microscopic semantics specific to each passage.

Open-Domain Question Answering Retrieval +1

Ensemble of Averages: Improving Model Selection and Boosting Performance in Domain Generalization

1 code implementation21 Oct 2021 Devansh Arpit, Huan Wang, Yingbo Zhou, Caiming Xiong

We first show that this chaotic behavior exists even along the training optimization trajectory of a single model, and propose a simple model averaging protocol that both significantly boosts domain generalization and diminishes the impact of stochasticity by improving the rank correlation between the in-domain validation accuracy and out-domain test accuracy, which is crucial for reliable early stopping.

Domain Generalization Model Selection

Long Document Summarization with Top-Down and Bottom-Up Representation Inference

no code implementations29 Sep 2021 Bo Pang, Erik Nijkamp, Wojciech Maciej Kryscinski, Silvio Savarese, Yingbo Zhou, Caiming Xiong

Critical to the success of a summarization model is the faithful inference of latent representations of words or tokens in the source documents.

Document Summarization

RnG-KBQA: Generation Augmented Iterative Ranking for Knowledge Base Question Answering

1 code implementation ACL 2022 Xi Ye, Semih Yavuz, Kazuma Hashimoto, Yingbo Zhou, Caiming Xiong

We present RnG-KBQA, a Rank-and-Generate approach for KBQA, which remedies the coverage issue with a generation model while preserving a strong generalization capability.

Entity Linking Knowledge Base Question Answering +1

CoCo: Controllable Counterfactuals for Evaluating Dialogue State Trackers

2 code implementations ICLR 2021 Shiyang Li, Semih Yavuz, Kazuma Hashimoto, Jia Li, Tong Niu, Nazneen Rajani, Xifeng Yan, Yingbo Zhou, Caiming Xiong

Dialogue state trackers have made significant progress on benchmark datasets, but their generalization capability to novel and realistic scenarios beyond the held-out conversations is less understood.

Ranked #2 on Multi-domain Dialogue State Tracking on MULTIWOZ 2.1 (using extra training data)

counterfactual Dialogue State Tracking +1

Unsupervised Paraphrasing with Pretrained Language Models

no code implementations EMNLP 2021 Tong Niu, Semih Yavuz, Yingbo Zhou, Nitish Shirish Keskar, Huan Wang, Caiming Xiong

To enforce a surface form dissimilar from the input, whenever the language model emits a token contained in the source sequence, DB prevents the model from outputting the subsequent source token for the next generation step.

Blocking Language Modelling +3

Online Structured Meta-learning

no code implementations NeurIPS 2020 Huaxiu Yao, Yingbo Zhou, Mehrdad Mahdavi, Zhenhui Li, Richard Socher, Caiming Xiong

When a new task is encountered, it constructs a meta-knowledge pathway by either utilizing the most relevant knowledge blocks or exploring new blocks.

Meta-Learning

Representation Learning for Sequence Data with Deep Autoencoding Predictive Components

2 code implementations ICLR 2021 Junwen Bai, Weiran Wang, Yingbo Zhou, Caiming Xiong

We propose Deep Autoencoding Predictive Components (DAPC) -- a self-supervised representation learning method for sequence data, based on the intuition that useful representations of sequence data should exhibit a simple structure in the latent space.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +3

WOAD: Weakly Supervised Online Action Detection in Untrimmed Videos

no code implementations CVPR 2021 Mingfei Gao, Yingbo Zhou, ran Xu, Richard Socher, Caiming Xiong

Online action detection in untrimmed videos aims to identify an action as it happens, which makes it very important for real-time applications.

Action Recognition Online Action Detection

Fast and Robust Unsupervised Contextual Biasing for Speech Recognition

no code implementations4 May 2020 Young Mo Kang, Yingbo Zhou

A common framework is to dynamically construct a small language model from the provided contextual mini corpus and interpolate its score with the main language model during the decoding process.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +2

An investigation of phone-based subword units for end-to-end speech recognition

no code implementations8 Apr 2020 Weiran Wang, Guangsen Wang, Aadyot Bhatnagar, Yingbo Zhou, Caiming Xiong, Richard Socher

For Switchboard, our phone-based BPE system achieves 6. 8\%/14. 4\% word error rate (WER) on the Switchboard/CallHome portion of the test set while joint decoding achieves 6. 3\%/13. 3\% WER.

Language Modelling speech-recognition +1

Differentially Private Deep Learning with Smooth Sensitivity

no code implementations1 Mar 2020 Lichao Sun, Yingbo Zhou, Philip S. Yu, Caiming Xiong

Ensuring the privacy of sensitive data used to train modern machine learning models is of paramount importance in many areas of practice.

Near-Zero-Cost Differentially Private Deep Learning with Teacher Ensembles

no code implementations25 Sep 2019 Lichao Sun, Yingbo Zhou, Jia Li, Richard Socher, Philip S. Yu, Caiming Xiong

Ensuring the privacy of sensitive data used to train modern machine learning models is of paramount importance in many areas of practice.

Private Deep Learning with Teacher Ensembles

no code implementations5 Jun 2019 Lichao Sun, Yingbo Zhou, Ji Wang, Jia Li, Richard Sochar, Philip S. Yu, Caiming Xiong

Privacy-preserving deep learning is crucial for deploying deep neural network based solutions, especially when the model works on data that contains sensitive information.

Ensemble Learning Knowledge Distillation +2

Learn to Grow: A Continual Structure Learning Framework for Overcoming Catastrophic Forgetting

no code implementations31 Mar 2019 Xilai Li, Yingbo Zhou, Tianfu Wu, Richard Socher, Caiming Xiong

Addressing catastrophic forgetting is one of the key challenges in continual learning where machine learning systems are trained with sequential or streaming tasks.

Continual Learning Neural Architecture Search +1

Augmented Cyclic Adversarial Learning for Low Resource Domain Adaptation

2 code implementations ICLR 2019 Ehsan Hosseini-Asl, Yingbo Zhou, Caiming Xiong, Richard Socher

In low-resource supervised setting, the results show that our approach improves absolute performance by 14% and 4% when adapting SVHN to MNIST and vice versa, respectively, which outperforms unsupervised domain adaptation methods that require high-resource unlabeled target domain.

speech-recognition Speech Recognition +1

Improved Regularization Techniques for End-to-End Speech Recognition

no code implementations19 Dec 2017 Yingbo Zhou, Caiming Xiong, Richard Socher

We augment audio data through random perturbations of tempo, pitch, volume, temporal alignment, and adding random noise. We further investigate the effect of dropout when applied to the inputs of all layers of the network.

Data Augmentation speech-recognition +1

Improving End-to-End Speech Recognition with Policy Learning

no code implementations19 Dec 2017 Yingbo Zhou, Caiming Xiong, Richard Socher

However, there is usually a disparity between the negative maximum likelihood and the performance metric used in speech recognition, e. g., word error rate (WER).

speech-recognition Speech Recognition

Spelling Correction as a Foreign Language

no code implementations21 May 2017 Yingbo Zhou, Utkarsh Porwal, Roberto Konow

In this paper, we reformulated the spell correction problem as a machine translation task under the encoder-decoder framework.

Feature Engineering Language Modelling +3

On Optimality Conditions for Auto-Encoder Signal Recovery

no code implementations ICLR 2018 Devansh Arpit, Yingbo Zhou, Hung Q. Ngo, Nils Napp, Venu Govindaraju

Auto-Encoders are unsupervised models that aim to learn patterns from observed data by minimizing a reconstruction cost.

Normalization Propagation: A Parametric Technique for Removing Internal Covariate Shift in Deep Networks

no code implementations4 Mar 2016 Devansh Arpit, Yingbo Zhou, Bhargava U. Kota, Venu Govindaraju

While the authors of Batch Normalization (BN) identify and address an important problem involved in training deep networks-- Internal Covariate Shift-- the current solution has certain drawbacks.

Maximum Entropy Binary Encoding for Face Template Protection

no code implementations5 Dec 2015 Rohit Kumar Pandey, Yingbo Zhou, Bhargava Urala Kota, Venu Govindaraju

In this paper we present a framework for secure identification using deep neural networks, and apply it to the task of template protection for face authentication.

Deep Secure Encoding: An Application to Face Recognition

no code implementations14 Jun 2015 Rohit Pandey, Yingbo Zhou, Venu Govindaraju

In this paper we present Deep Secure Encoding: a framework for secure classification using deep neural networks, and apply it to the task of biometric template protection for faces.

Face Recognition General Classification

Why Regularized Auto-Encoders learn Sparse Representation?

no code implementations21 May 2015 Devansh Arpit, Yingbo Zhou, Hung Ngo, Venu Govindaraju

While the authors of Batch Normalization (BN) identify and address an important problem involved in training deep networks-- \textit{Internal Covariate Shift}-- the current solution has certain drawbacks.

Parallel Feature Selection Inspired by Group Testing

no code implementations NeurIPS 2014 Yingbo Zhou, Utkarsh Porwal, Ce Zhang, Hung Q. Ngo, XuanLong Nguyen, Christopher Ré, Venu Govindaraju

Superior performance of our method is demonstrated on a challenging relation extraction task from a very large data set that have both redundant features and sample size in the order of millions.

feature selection General Classification +1

Is Joint Training Better for Deep Auto-Encoders?

no code implementations6 May 2014 Yingbo Zhou, Devansh Arpit, Ifeoma Nwogu, Venu Govindaraju

But due to the greedy scheme of the layerwise training technique, the parameters of lower layers are fixed when training higher layers.

Cannot find the paper you are looking for? You can Submit a new open access paper.