Search Results for author: BoWen Zhou

Found 45 papers, 18 papers with code

Learn to Copy from the Copying History: Correlational Copy Network for Abstractive Summarization

1 code implementation EMNLP 2021 Haoran Li, Song Xu, Peng Yuan, Yujia Wang, Youzheng Wu, Xiaodong He, BoWen Zhou

It thereby takes advantage of prior copying distributions and, at each time step, explicitly encourages the model to copy the input word that is relevant to the previously copied one.

Ranked #11 on Abstractive Text Summarization on CNN / Daily Mail (using extra training data)

Abstractive Text Summarization News Summarization

On Large Language Models' Hallucination with Regard to Known Facts

no code implementations29 Mar 2024 Che Jiang, Biqing Qi, Xiangyu Hong, Dayuan Fu, Yang Cheng, Fandong Meng, Mo Yu, BoWen Zhou, Jie zhou

In hallucinated cases, the output token's information rarely demonstrates abrupt increases and consistent superiority in the later stages of the model.

Hallucination

Mastering Text, Code and Math Simultaneously via Fusing Highly Specialized Language Models

no code implementations13 Mar 2024 Ning Ding, Yulin Chen, Ganqu Cui, Xingtai Lv, Weilin Zhao, Ruobing Xie, BoWen Zhou, Zhiyuan Liu, Maosong Sun

Underlying data distributions of natural language, programming code, and mathematical symbols vary vastly, presenting a complex challenge for large language models (LLMs) that strive to achieve high performance across all three domains simultaneously.

Math

Contrastive Augmented Graph2Graph Memory Interaction for Few Shot Continual Learning

no code implementations7 Mar 2024 Biqing Qi, Junqi Gao, Xingquan Chen, Dong Li, Jianxing Liu, Ligang Wu, BoWen Zhou

However, current EM-based methods retrieves memory globally by performing Vector-to-Vector (V2V) interaction between features corresponding to the input and prototypes stored in EM, neglecting the geometric structure of local features.

Few-Shot Class-Incremental Learning Few-Shot Learning +1

CoGenesis: A Framework Collaborating Large and Small Language Models for Secure Context-Aware Instruction Following

no code implementations5 Mar 2024 Kaiyan Zhang, Jianyu Wang, Ermo Hua, Biqing Qi, Ning Ding, BoWen Zhou

With the advancement of language models (LMs), their exposure to private data is increasingly inevitable, and their deployment (especially for smaller ones) on personal devices, such as PCs and smartphones, has become a prevailing trend.

Instruction Following

Interactive Continual Learning: Fast and Slow Thinking

1 code implementation5 Mar 2024 Biqing Qi, Xingquan Chen, Junqi Gao, Dong Li, Jianxing Liu, Ligang Wu, BoWen Zhou

Drawing on Complementary Learning System theory, this paper presents a novel Interactive Continual Learning (ICL) framework, enabled by collaborative interactions among models of various sizes.

Continual Learning Outlier Detection +1

Investigating Deep Watermark Security: An Adversarial Transferability Perspective

no code implementations26 Feb 2024 Biqing Qi, Junqi Gao, Yiang Luo, Jianxing Liu, Ligang Wu, BoWen Zhou

The rise of generative neural networks has triggered an increased demand for intellectual property (IP) protection in generated content.

Critical Data Size of Language Models from a Grokking Perspective

no code implementations19 Jan 2024 Xuekai Zhu, Yao Fu, BoWen Zhou, Zhouhan Lin

We formalize the phase transition under the grokking configuration into the Data Efficiency Hypothesis and identify data insufficiency, sufficiency, and surplus regimes in language models training dynamics.

Language Modelling Memorization

Generative Multi-Modal Knowledge Retrieval with Large Language Models

no code implementations16 Jan 2024 Xinwei Long, Jiali Zeng, Fandong Meng, Zhiyuan Ma, Kaiyan Zhang, BoWen Zhou, Jie zhou

Knowledge retrieval with multi-modal queries plays a crucial role in supporting knowledge-intensive multi-modal applications.

Retrieval

AdapEdit: Spatio-Temporal Guided Adaptive Editing Algorithm for Text-Based Continuity-Sensitive Image Editing

1 code implementation13 Dec 2023 Zhiyuan Ma, Guoli Jia, BoWen Zhou

With the great success of text-conditioned diffusion models in creative text-to-image generation, various text-driven image editing approaches have attracted the attentions of many researchers.

Text-to-Image Generation

LMD: Faster Image Reconstruction with Latent Masking Diffusion

1 code implementation13 Dec 2023 Zhiyuan Ma, zhihuan yu, Jianjun Li, BoWen Zhou

Then, we combine the advantages of MAEs and DPMs to design a progressive masking diffusion model, which gradually increases the masking proportion by three different schedulers and reconstructs the latent features from simple to difficult, without sequentially performing denoising diffusion as in DPMs or using fixed high masking ratio as in MAEs, so as to alleviate the high training time-consumption predicament.

Denoising Image Reconstruction

Large Model Based Referring Camouflaged Object Detection

no code implementations28 Nov 2023 Shupeng Cheng, Ge-Peng Ji, Pengda Qin, Deng-Ping Fan, BoWen Zhou, Peng Xu

Our motivation is to make full use of the semantic intelligence and intrinsic knowledge of recent Multimodal Large Language Models (MLLMs) to decompose this complex task in a human-like way.

Object object-detection +2

Sparse Low-rank Adaptation of Pre-trained Language Models

1 code implementation20 Nov 2023 Ning Ding, Xingtai Lv, Qiaosen Wang, Yulin Chen, BoWen Zhou, Zhiyuan Liu, Maosong Sun

Recognizing the need for more flexible adaptation, we extend the methodology of LoRA to an innovative approach we call sparse low-rank adaptation (SoRA) that enables dynamic adjustments to the intrinsic rank during the adaptation process.

Memorization

Large Language Models are Zero Shot Hypothesis Proposers

no code implementations10 Nov 2023 Biqing Qi, Kaiyan Zhang, Haoxiang Li, Kai Tian, Sihang Zeng, Zhang-Ren Chen, BoWen Zhou

We subsequently evaluate the hypothesis generation capabilities of various top-tier instructed models in zero-shot, few-shot, and fine-tuning settings, including both closed and open-source LLMs.

CRaSh: Clustering, Removing, and Sharing Enhance Fine-tuning without Full Large Language Model

no code implementations24 Oct 2023 Kaiyan Zhang, Ning Ding, Biqing Qi, Xuekai Zhu, Xinwei Long, BoWen Zhou

Instruction tuning has recently been recognized as an effective way of aligning Large Language Models (LLMs) to enhance their generalization ability across various tasks.

Clustering Language Modelling +1

When Super-Resolution Meets Camouflaged Object Detection: A Comparison Study

no code implementations8 Aug 2023 Juan Wen, Shupeng Cheng, Peng Xu, BoWen Zhou, Radu Timofte, Weiyan Hou, Luc van Gool

Super Resolution (SR) and Camouflaged Object Detection (COD) are two hot topics in computer vision with various joint applications.

Object object-detection +2

Enhancing Chat Language Models by Scaling High-quality Instructional Conversations

1 code implementation23 May 2023 Ning Ding, Yulin Chen, Bokai Xu, Yujia Qin, Zhi Zheng, Shengding Hu, Zhiyuan Liu, Maosong Sun, BoWen Zhou

Fine-tuning on instruction data has been widely validated as an effective practice for implementing chat language models like ChatGPT.

PaD: Program-aided Distillation Can Teach Small Models Reasoning Better than Chain-of-thought Fine-tuning

1 code implementation23 May 2023 Xuekai Zhu, Biqing Qi, Kaiyan Zhang, Xinwei Long, Zhouhan Lin, BoWen Zhou

While large language models (LLMs) excel in various natural language processing tasks, their huge size and the inaccessibility of parameters present challenges for practical deployment.

Arithmetic Reasoning GSM8K +1

SAM Struggles in Concealed Scenes -- Empirical Study on "Segment Anything"

no code implementations12 Apr 2023 Ge-Peng Ji, Deng-Ping Fan, Peng Xu, Ming-Ming Cheng, BoWen Zhou, Luc van Gool

Segmenting anything is a ground-breaking step toward artificial general intelligence, and the Segment Anything Model (SAM) greatly fosters the foundation models for computer vision.

PetsGAN: Rethinking Priors for Single Image Generation

2 code implementations3 Mar 2022 ZiCheng Zhang, Yinglu Liu, Congying Han, Hailin Shi, Tiande Guo, BoWen Zhou

Moreover, we apply our method to other image manipulation tasks (e. g., style transfer, harmonization), and the results further prove the effectiveness and efficiency of our method.

Image Generation Image Manipulation +2

On the Convergence and Robustness of Adversarial Training

no code implementations15 Dec 2021 Yisen Wang, Xingjun Ma, James Bailey, JinFeng Yi, BoWen Zhou, Quanquan Gu

In this paper, we propose such a criterion, namely First-Order Stationary Condition for constrained optimization (FOSC), to quantitatively evaluate the convergence quality of adversarial examples found in the inner maximization.

Trustworthy AI: From Principles to Practices

no code implementations4 Oct 2021 Bo Li, Peng Qi, Bo Liu, Shuai Di, Jingen Liu, JiQuan Pei, JinFeng Yi, BoWen Zhou

In this review, we provide AI practitioners with a comprehensive guide for building trustworthy AI systems.

Fairness

The JDDC 2.0 Corpus: A Large-Scale Multimodal Multi-Turn Chinese Dialogue Dataset for E-commerce Customer Service

no code implementations27 Sep 2021 Nan Zhao, Haoran Li, Youzheng Wu, Xiaodong He, BoWen Zhou

We present the solutions of top-5 teams participating in the JDDC multimodal dialogue challenge based on this dataset, which provides valuable insights for further researches on the multimodal dialogue task.

CUSTOM: Aspect-Oriented Product Summarization for E-Commerce

1 code implementation18 Aug 2021 Jiahui Liang, Junwei Bao, Yifan Wang, Youzheng Wu, Xiaodong He, BoWen Zhou

To address the problem, we propose CUSTOM, aspect-oriented product summarization for e-commerce, which generates diverse and controllable summaries towards different product aspects.

Augmentation Pathways Network for Visual Recognition

1 code implementation26 Jul 2021 Yalong Bai, Mohan Zhou, Wei zhang, BoWen Zhou, Tao Mei

Experimental results on ImageNet demonstrate the compatibility and effectiveness on a much wider range of augmentations, while consuming fewer parameters and lower computational costs at inference time.

Data Augmentation

Open Temporal Relation Extraction for Question Answering

no code implementations AKBC 2021 Chao Shang, Peng Qi, Guangtao Wang, Jing Huang, Youzheng Wu, BoWen Zhou

Understanding the temporal relations among events in text is a critical aspect of reading comprehension, which can be evaluated in the form of temporal question answering (TQA).

Question Answering Reading Comprehension +2

Joint System-Wise Optimization for Pipeline Goal-Oriented Dialog System

no code implementations9 Jun 2021 Zichuan Lin, Jing Huang, BoWen Zhou, Xiaodong He, Tengyu Ma

Recent work (Takanobu et al., 2020) proposed the system-wise evaluation on dialog systems and found that improvement on individual components (e. g., NLU, policy) in prior work may not necessarily bring benefit to pipeline systems in system-wise evaluation.

Data Augmentation Goal-Oriented Dialog

Conversational AI Systems for Social Good: Opportunities and Challenges

no code implementations13 May 2021 Peng Qi, Jing Huang, Youzheng Wu, Xiaodong He, BoWen Zhou

Conversational artificial intelligence (ConvAI) systems have attracted much academic and commercial attention recently, making significant progress on both fronts.

SGG: Learning to Select, Guide, and Generate for Keyphrase Generation

1 code implementation NAACL 2021 Jing Zhao, Junwei Bao, Yifan Wang, Youzheng Wu, Xiaodong He, BoWen Zhou

Keyphrases, that concisely summarize the high-level topics discussed in a document, can be categorized into present keyphrase which explicitly appears in the source text, and absent keyphrase which does not match any contiguous subsequence but is highly semantically related to the source.

Keyphrase Generation Text Generation

K-PLUG: Knowledge-injected Pre-trained Language Model for Natural Language Understanding and Generation in E-Commerce

1 code implementation Findings (EMNLP) 2021 Song Xu, Haoran Li, Peng Yuan, Yujia Wang, Youzheng Wu, Xiaodong He, Ying Liu, BoWen Zhou

K-PLUG achieves new state-of-the-art results on a suite of domain-specific NLP tasks, including product knowledge base completion, abstractive product summarization, and multi-turn dialogue, significantly outperforms baselines across the board, which demonstrates that the proposed method effectively learns a diverse set of domain-specific knowledge for both language understanding and generation tasks.

Knowledge Base Completion Language Modelling +2

Graph Ensemble Learning over Multiple Dependency Trees for Aspect-level Sentiment Classification

no code implementations NAACL 2021 Xiaochen Hou, Peng Qi, Guangtao Wang, Rex Ying, Jing Huang, Xiaodong He, BoWen Zhou

Recent work on aspect-level sentiment classification has demonstrated the efficacy of incorporating syntactic structures such as dependency trees with graph neural networks(GNN), but these approaches are usually vulnerable to parsing errors.

Ensemble Learning General Classification +2

HAIR: Head-mounted AR Intention Recognition

no code implementations22 Feb 2021 David Puljiz, BoWen Zhou, Ke Ma, Björn Hein

In this paper we propose an intention recognition system that is based purely on a portable head-mounted display.

Intent Detection Robotics Human-Computer Interaction

Conversational Query Rewriting with Self-supervised Learning

no code implementations9 Feb 2021 Hang Liu, Meng Chen, Youzheng Wu, Xiaodong He, BoWen Zhou

Conversational Query Rewriting (CQR) aims to simplify the multi-turn dialogue modeling into a single-turn problem by explicitly rewriting the conversational query into a self-contained utterance.

Self-Supervised Learning

K-PLUG: KNOWLEDGE-INJECTED PRE-TRAINED LANGUAGE MODEL FOR NATURAL LANGUAGE UNDERSTANDING AND GENERATION

1 code implementation1 Jan 2021 Song Xu, Haoran Li, Peng Yuan, Yujia Wang, Youzheng Wu, Xiaodong He, Ying Liu, BoWen Zhou

K-PLUG achieves new state-of-the-art results on a suite of domain-specific NLP tasks, including product knowledge base completion, abstractive product summarization, and multi-turn dialogue, significantly outperforms baselines across the board, which demonstrates that the proposed method effectively learns a diverse set of domain-specific knowledge for both language understanding and generation tasks.

Chatbot Knowledge Base Completion +4

Improving Prosody Modelling with Cross-Utterance BERT Embeddings for End-to-end Speech Synthesis

no code implementations6 Nov 2020 Guanghui Xu, Wei Song, Zhengchen Zhang, Chao Zhang, Xiaodong He, BoWen Zhou

Despite prosody is related to the linguistic information up to the discourse structure, most text-to-speech (TTS) systems only take into account that within each sentence, which makes it challenging when converting a paragraph of texts into natural and expressive speech.

Sentence Sentence Embeddings +1

Neural Kalman Filtering for Speech Enhancement

no code implementations28 Jul 2020 Wei Xue, Gang Quan, Chao Zhang, Guohong Ding, Xiaodong He, BoWen Zhou

Statistical signal processing based speech enhancement methods adopt expert knowledge to design the statistical models and linear filters, which is complementary to the deep neural network (DNN) based methods which are data-driven.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +2

Incremental Learning for End-to-End Automatic Speech Recognition

no code implementations11 May 2020 Li Fu, Xiaoxiao Li, Libo Zi, Zhengchen Zhang, Youzheng Wu, Xiaodong He, BoWen Zhou

In this paper, we propose an incremental learning method for end-to-end Automatic Speech Recognition (ASR) which enables an ASR system to perform well on new tasks while maintaining the performance on its originally learned ones.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +3

Finding Needle in a Million Metrics: Anomaly Detection in a Large-scale Computational Advertising Platform

no code implementations23 Feb 2016 Bowen Zhou, Shahriar Shariat

Online media offers opportunities to marketers to deliver brand messages to a large audience.

Anomaly Detection

Cannot find the paper you are looking for? You can Submit a new open access paper.