Search Results for author: Rui Yan

Found 193 papers, 77 papers with code

Fast Low-rank Representation based Spatial Pyramid Matching for Image Classification

no code implementations22 Sep 2014 Xi Peng, Rui Yan, Bo Zhao, Huajin Tang, Zhang Yi

Although the methods achieve a higher recognition rate than the traditional SPM, they consume more time to encode the local descriptors extracted from the image.

General Classification Image Classification +1

Automatic Subspace Learning via Principal Coefficients Embedding

no code implementations17 Nov 2014 Xi Peng, Jiwen Lu, Zhang Yi, Rui Yan

In this paper, we address two challenging problems in unsupervised subspace learning: 1) how to automatically identify the feature dimension of the learned subspace (i. e., automatic subspace learning), and 2) how to learn the underlying subspace in the presence of Gaussian noise (i. e., robust subspace learning).

Backward and Forward Language Modeling for Constrained Sentence Generation

no code implementations21 Dec 2015 Lili Mou, Rui Yan, Ge Li, Lu Zhang, Zhi Jin

Provided a specific word, we use RNNs to generate previous words and future words, either simultaneously or asynchronously, resulting in two model variants.

Language Modelling Machine Translation +4

How Transferable are Neural Networks in NLP Applications?

no code implementations EMNLP 2016 Lili Mou, Zhao Meng, Rui Yan, Ge Li, Yan Xu, Lu Zhang, Zhi Jin

Transfer learning is aimed to make use of valuable knowledge in a source domain to help model performance in a target domain.

Transfer Learning

StalemateBreaker: A Proactive Content-Introducing Approach to Automatic Human-Computer Conversation

no code implementations15 Apr 2016 Xiang Li, Lili Mou, Rui Yan, Ming Zhang

In this paper, we propose StalemateBreaker, a conversation system that can proactively introduce new content when appropriate.

Dialogue Session Segmentation by Embedding-Enhanced TextTiling

no code implementations13 Oct 2016 Yiping Song, Lili Mou, Rui Yan, Li Yi, Zinan Zhu, Xiaohua Hu, Ming Zhang

In human-computer conversation systems, the context of a user-issued utterance is particularly important because it provides useful background information of the conversation.

Word Embeddings

Two are Better than One: An Ensemble of Retrieval- and Generation-Based Dialog Systems

2 code implementations23 Oct 2016 Yiping Song, Rui Yan, Xiang Li, Dongyan Zhao, Ming Zhang

In this paper, we propose a novel ensemble of retrieval-based and generation-based dialog systems in the open domain.

Retrieval

Neural Emoji Recommendation in Dialogue Systems

no code implementations14 Dec 2016 Ruobing Xie, Zhiyuan Liu, Rui Yan, Maosong Sun

It indicates that our method could well capture the contextual information and emotion flow in dialogues, which is significant for emoji recommendation.

General Classification

RUBER: An Unsupervised Method for Automatic Evaluation of Open-Domain Dialog Systems

1 code implementation11 Jan 2017 Chongyang Tao, Lili Mou, Dongyan Zhao, Rui Yan

Open-domain human-computer conversation has been attracting increasing attention over the past few years.

Dialogue Evaluation Retrieval

A Constrained Sequence-to-Sequence Neural Model for Sentence Simplification

no code implementations7 Apr 2017 Yaoyuan Zhang, Zhenxu Ye, Yansong Feng, Dongyan Zhao, Rui Yan

For word-level studies, words are simplified but also have potential grammar errors due to different usages of words before and after simplification.

Sentence

Diversifying Neural Conversation Model with Maximal Marginal Relevance

no code implementations IJCNLP 2017 Yiping Song, Zhiliang Tian, Dongyan Zhao, Ming Zhang, Rui Yan

However, traditional seq2seq suffer from a severe weakness: during beam search decoding, they tend to rank universal replies at the top of the candidate list, resulting in the lack of diversity among candidate replies.

Document Summarization Information Retrieval +1

Style Transfer in Text: Exploration and Evaluation

2 code implementations18 Nov 2017 Zhenxin Fu, Xiaoye Tan, Nanyun Peng, Dongyan Zhao, Rui Yan

Results show that the proposed content preservation metric is highly correlate to human judgments, and the proposed models are able to generate sentences with higher style transfer strength and similar content preservation score comparing to auto-encoder.

Style Transfer Text Style Transfer

Scale Up Event Extraction Learning via Automatic Training Data Generation

no code implementations11 Dec 2017 Ying Zeng, Yansong Feng, Rong Ma, Zheng Wang, Rui Yan, Chongde Shi, Dongyan Zhao

We show that this large volume of training data not only leads to a better event extractor, but also allows us to detect multiple typed events.

Event Extraction

Tree2Tree Learning with Memory Unit

no code implementations ICLR 2018 Ning Miao, Hengliang Wang, Ran Le, Chongyang Tao, Mingyue Shang, Rui Yan, Dongyan Zhao

Traditional recurrent neural network (RNN) or convolutional neural net- work (CNN) based sequence-to-sequence model can not handle tree structural data well.

Machine Translation Translation

Topic-Based Question Generation

no code implementations ICLR 2018 Wenpeng Hu, Bing Liu, Rui Yan, Dongyan Zhao, Jinwen Ma

In the paper, we propose a new question generation problem, which also requires the input of a target topic in addition to a piece of descriptive text.

Chatbot Descriptive +3

Marrying up Regular Expressions with Neural Networks: A Case Study for Spoken Language Understanding

no code implementations ACL 2018 Bingfeng Luo, Yansong Feng, Zheng Wang, Songfang Huang, Rui Yan, Dongyan Zhao

The success of many natural language processing (NLP) tasks is bound by the number and quality of annotated data, but there is often a shortage of such training data.

Intent Detection slot-filling +2

Improving Matching Models with Hierarchical Contextualized Representations for Multi-turn Response Selection

no code implementations22 Aug 2018 Chongyang Tao, Wei Wu, Can Xu, Yansong Feng, Dongyan Zhao, Rui Yan

In this paper, we study context-response matching with pre-trained contextualized representations for multi-turn response selection in retrieval-based chatbots.

Dialogue Generation Retrieval +1

Revisiting Random Binning Features: Fast Convergence and Strong Parallelizability

2 code implementations14 Sep 2018 Lingfei Wu, Ian E. H. Yen, Jie Chen, Rui Yan

We thus propose the first analysis of RB from the perspective of optimization, which by interpreting RB as a Randomized Block Coordinate Descent in the infinite-dimensional space, gives a faster convergence rate compared to that of other random features.

Iterative Document Representation Learning Towards Summarization with Polishing

1 code implementation EMNLP 2018 Xiuying Chen, Shen Gao, Chongyang Tao, Yan Song, Dongyan Zhao, Rui Yan

In this paper, we introduce Iterative Text Summarization (ITS), an iteration-based model for supervised extractive text summarization, inspired by the observation that it is often necessary for a human to read an article multiple times in order to fully understand and summarize its contents.

Extractive Text Summarization Representation Learning +1

On the Abstractiveness of Neural Document Summarization

no code implementations EMNLP 2018 Fangfang Zhang, Jin-Ge Yao, Rui Yan

Many modern neural document summarization systems based on encoder-decoder networks are designed to produce abstractive summaries.

Abstractive Text Summarization Document Summarization

Plan-And-Write: Towards Better Automatic Storytelling

2 code implementations14 Nov 2018 Lili Yao, Nanyun Peng, Ralph Weischedel, Kevin Knight, Dongyan Zhao, Rui Yan

Automatic storytelling is challenging since it requires generating long, coherent natural language to describes a sensible sequence of events.

Story Generation

CGMH: Constrained Sentence Generation by Metropolis-Hastings Sampling

1 code implementation14 Nov 2018 Ning Miao, Hao Zhou, Lili Mou, Rui Yan, Lei LI

In real-world applications of natural language generation, there are often constraints on the target sentences in addition to fluency and naturalness requirements.

Sentence Text Generation

Chat More If You Like: Dynamic Cue Words Planning to Flow Longer Conversations

no code implementations19 Nov 2018 Lili Yao, Ruijian Xu, Chao Li, Dongyan Zhao, Rui Yan

To build an open-domain multi-turn conversation system is one of the most interesting and challenging tasks in Artificial Intelligence.

Abstractive Text Summarization by Incorporating Reader Comments

no code implementations13 Dec 2018 Shen Gao, Xiuying Chen, Piji Li, Zhaochun Ren, Lidong Bing, Dongyan Zhao, Rui Yan

To tackle this problem, we propose the task of reader-aware abstractive summary generation, which utilizes the reader comments to help the model produce better summary about the main aspect.

Reader-Aware Summarization

Find a Reasonable Ending for Stories: Does Logic Relation Help the Story Cloze Test?

no code implementations13 Dec 2018 Mingyue Shang, Zhenxin Fu, Hongzhi Yin, Bo Tang, Dongyan Zhao, Rui Yan

In this paper, we incorporate the logic information with the help of the Natural Language Inference (NLI) task to the Story Cloze Test (SCT).

Cloze Test Natural Language Inference +2

Product-Aware Answer Generation in E-Commerce Question-Answering

1 code implementation23 Jan 2019 Shen Gao, Zhaochun Ren, Yihong Eric Zhao, Dongyan Zhao, Dawei Yin, Rui Yan

In this paper, we propose the task of product-aware answer generation, which tends to generate an accurate and complete answer from large-scale unlabeled e-commerce reviews and product attributes.

Answer Generation Question Answering

GSN: A Graph-Structured Network for Multi-Party Dialogues

1 code implementation31 May 2019 Wenpeng Hu, Zhangming Chan, Bing Liu, Dongyan Zhao, Jinwen Ma, Rui Yan

Existing neural models for dialogue response generation assume that utterances are sequentially organized.

Response Generation

A Document-grounded Matching Network for Response Selection in Retrieval-based Chatbots

no code implementations11 Jun 2019 Xueliang Zhao, Chongyang Tao, Wei Wu, Can Xu, Dongyan Zhao, Rui Yan

We present a document-grounded matching network (DGMN) for response selection that can power a knowledge-aware retrieval-based chatbot system.

Chatbot Retrieval

Mimicking Human Process: Text Representation via Latent Semantic Clustering for Classification

no code implementations18 Jun 2019 Xiaoye Tan, Rui Yan, Chongyang Tao, Mingrui Wu

Considering that words with different characteristic in the text have different importance for classification, grouping them together separately can strengthen the semantic expression of each part.

Classification Clustering +1

Are Training Samples Correlated? Learning to Generate Dialogue Responses with Multiple References

no code implementations ACL 2019 Lisong Qiu, Juntao Li, Wei Bi, Dongyan Zhao, Rui Yan

Due to its potential applications, open-domain dialogue generation has become popular and achieved remarkable progress in recent years, but sometimes suffers from generic responses.

Dialogue Generation valid

Learning towards Abstractive Timeline Summarization

1 code implementation IJCAI 2019 2019 Xiuying Chen, Zhangming Chan, Shen Gao, Meng-Hsuan Yu, Dongyan Zhao, Rui Yan

Timeline summarization targets at concisely summarizing the evolution trajectory along the timeline and existing timeline summarization approaches are all based on extractive methods. In this paper, we propose the task of abstractive timeline summarization, which tends to concisely paraphrase the information in the time-stamped events. Unlike traditional document summarization, timeline summarization needs to model the time series information of the input events and summarize important events in chronological order. To tackle this challenge, we propose a memory-based timeline summarization model (MTS). Concretely, we propose a time-event memory to establish a timeline, and use the time position of events on this timeline to guide generation process. Besides, in each decoding step, we incorporate event-level information into word-level attention to avoid confusion between events. Extensive experiments are conducted on a large-scale real-world dataset, and the results show that MTS achieves the state-of-the-art performance in terms of both automatic and human evaluations.

Document Summarization Timeline Summarization +2

Relation-Aware Entity Alignment for Heterogeneous Knowledge Graphs

1 code implementation22 Aug 2019 Yuting Wu, Xiao Liu, Yansong Feng, Zheng Wang, Rui Yan, Dongyan Zhao

Entity alignment is the task of linking entities with the same real-world identity from different knowledge graphs (KGs), which has been recently dominated by embedding-based methods.

Ranked #20 on Entity Alignment on DBP15k zh-en (using extra training data)

Entity Alignment Entity Embeddings +2

How to Write Summaries with Patterns? Learning towards Abstractive Summarization through Prototype Editing

1 code implementation IJCNLP 2019 Shen Gao, Xiuying Chen, Piji Li, Zhangming Chan, Dongyan Zhao, Rui Yan

There are two main challenges in this task: (1) the model needs to incorporate learned patterns from the prototype, but (2) should avoid copying contents other than the patternized words---such as irrelevant facts---into the generated summaries.

Abstractive Text Summarization

Learning from Positive and Unlabeled Data with Adversarial Training

no code implementations25 Sep 2019 Wenpeng Hu, Ran Le, Bing Liu, Feng Ji, Haiqing Chen, Dongyan Zhao, Jinwen Ma, Rui Yan

Positive-unlabeled (PU) learning learns a binary classifier using only positive and unlabeled examples without labeled negative examples.

Semi-supervised Text Style Transfer: Cross Projection in Latent Space

no code implementations IJCNLP 2019 Mingyue Shang, Piji Li, Zhenxin Fu, Lidong Bing, Dongyan Zhao, Shuming Shi, Rui Yan

Text style transfer task requires the model to transfer a sentence of one style to another style while retaining its original content meaning, which is a challenging problem that has long suffered from the shortage of parallel data.

Sentence Style Transfer +1

Multilingual Dialogue Generation with Shared-Private Memory

no code implementations6 Oct 2019 Chen Chen, Lisong Qiu, Zhenxin Fu, Dongyan Zhao, Junfei Liu, Rui Yan

Existing dialog systems are all monolingual, where features shared among different languages are rarely explored.

Cross-Lingual Transfer Dialogue Generation

RPM-Oriented Query Rewriting Framework for E-commerce Keyword-Based Sponsored Search

no code implementations28 Oct 2019 Xiuying Chen, Daorui Xiao, Shen Gao, Guojun Liu, Wei. Lin, Bo Zheng, Dongyan Zhao, Rui Yan

Sponsored search optimizes revenue and relevance, which is estimated by Revenue Per Mille (RPM).

Learning to Customize Model Structures for Few-shot Dialogue Generation Tasks

1 code implementation ACL 2020 Yiping Song, Zequn Liu, Wei Bi, Rui Yan, Ming Zhang

Training the generative models with minimal corpus is one of the critical challenges for building open-domain dialogue systems.

Dialogue Generation Language Modelling +1

Who Is Speaking to Whom? Learning to Identify Utterance Addressee in Multi-Party Conversations

no code implementations IJCNLP 2019 Ran Le, Wenpeng Hu, Mingyue Shang, Zhenjun You, Lidong Bing, Dongyan Zhao, Rui Yan

Previous research on dialogue systems generally focuses on the conversation between two participants, yet multi-party conversations which involve more than two participants within one session bring up a more complicated but realistic scenario.

Stick to the Facts: Learning towards a Fidelity-oriented E-Commerce Product Description Generation

no code implementations IJCNLP 2019 Zhangming Chan, Xiuying Chen, Yongliang Wang, Juntao Li, Zhiqiang Zhang, Kun Gai, Dongyan Zhao, Rui Yan

Different from other text generation tasks, in product description generation, it is of vital importance to generate faithful descriptions that stick to the product attribute information.

Attribute Text Generation

Query-bag Matching with Mutual Coverage for Information-seeking Conversations in E-commerce

1 code implementation7 Nov 2019 Zhenxin Fu, Feng Ji, Wenpeng Hu, Wei Zhou, Dongyan Zhao, Haiqing Chen, Rui Yan

Information-seeking conversation system aims at satisfying the information needs of users through conversations.

Text Matching

Low-Resource Knowledge-Grounded Dialogue Generation

no code implementations ICLR 2020 Xueliang Zhao, Wei Wu, Chongyang Tao, Can Xu, Dongyan Zhao, Rui Yan

In such a low-resource setting, we devise a disentangled response decoder in order to isolate parameters that depend on knowledge-grounded dialogues from the entire generation model.

Dialogue Generation Response Generation

Learning to Respond with Stickers: A Framework of Unifying Multi-Modality in Multi-Turn Dialog

1 code implementation10 Mar 2020 Shen Gao, Xiuying Chen, Chang Liu, Li Liu, Dongyan Zhao, Rui Yan

Stickers with vivid and engaging expressions are becoming increasingly popular in online messaging apps, and some works are dedicated to automatically select sticker response by matching text labels of stickers with previous utterances.

EnsembleGAN: Adversarial Learning for Retrieval-Generation Ensemble Model on Short-Text Conversation

no code implementations30 Apr 2020 Jiayi Zhang, Chongyang Tao, Zhenjing Xu, Qiaojing Xie, Wei Chen, Rui Yan

Aiming at generating responses that approximate the ground-truth and receive high ranking scores from the discriminator, the two generators learn to generate improved highly relevant responses and competitive unobserved candidates respectively, while the discriminative ranker is trained to identify true responses from adversarial ones, thus featuring the merits of both generator counterparts.

Language Modelling Retrieval +1

Epipolar Transformers

1 code implementation CVPR 2020 Yihui He, Rui Yan, Katerina Fragkiadaki, Shoou-I Yu

The intuition is: given a 2D location p in the current view, we would like to first find its corresponding point p' in a neighboring view, and then combine the features at p' with the features at p, thus leading to a 3D-aware feature at p. Inspired by stereo matching, the epipolar transformer leverages epipolar constraints and feature matching to approximate the features at p'.

2D Pose Estimation 3D Hand Pose Estimation +3

From Standard Summarization to New Tasks and Beyond: Summarization with Manifold Information

no code implementations10 May 2020 Shen Gao, Xiuying Chen, Zhaochun Ren, Dongyan Zhao, Rui Yan

Text summarization is the research area aiming at creating a short and condensed version of the original document, which conveys the main idea of the document in a few words.

Text Summarization

Cross-Lingual Low-Resource Set-to-Description Retrieval for Global E-Commerce

1 code implementation17 May 2020 Juntao Li, Chang Liu, Jian Wang, Lidong Bing, Hongsong Li, Xiaozhong Liu, Dongyan Zhao, Rui Yan

We manually collect a new and high-quality paired dataset, where each pair contains an unordered product attribute set in the source language and an informative product description in the target language.

Attribute Cross-Lingual Information Retrieval +1

Policy Evaluation and Seeking for Multi-Agent Reinforcement Learning via Best Response

no code implementations17 Jun 2020 Rui Yan, Xiaoming Duan, Zongying Shi, Yisheng Zhong, Jason R. Marden, Francesco Bullo

With this knowledge we propose a class of perturbed SBRD with the following property: only policies with maximum metric are observed with nonzero probability for a broad class of stochastic games with finite memory.

Multi-agent Reinforcement Learning reinforcement-learning +1

Social Adaptive Module for Weakly-supervised Group Activity Recognition

no code implementations ECCV 2020 Rui Yan, Lingxi Xie, Jinhui Tang, Xiangbo Shu, Qi Tian

This paper presents a new task named weakly-supervised group activity recognition (GAR) which differs from conventional GAR tasks in that only video-level labels are available, yet the important persons within each frame are not provided even in the training data.

Group Activity Recognition

Learning an Effective Context-Response Matching Model with Self-Supervised Tasks for Retrieval-based Dialogues

no code implementations14 Sep 2020 Ruijian Xu, Chongyang Tao, Daxin Jiang, Xueliang Zhao, Dongyan Zhao, Rui Yan

To address these issues, in this paper, we propose learning a context-response matching model with auxiliary self-supervised tasks designed for the dialogue data based on pre-trained language models.

Conversational Response Selection Retrieval

VMSMO: Learning to Generate Multimodal Summary for Video-based News Articles

1 code implementation EMNLP 2020 Mingzhe Li, Xiuying Chen, Shen Gao, Zhangming Chan, Dongyan Zhao, Rui Yan

Hence, in this paper, we propose the task of Video-based Multimodal Summarization with Multimodal Output (VMSMO) to tackle such a problem.

Learning to Respond with Your Favorite Stickers: A Framework of Unifying Multi-Modality and User Preference in Multi-Turn Dialog

no code implementations5 Nov 2020 Shen Gao, Xiuying Chen, Li Liu, Dongyan Zhao, Rui Yan

Hence, in this paper, we propose to recommend an appropriate sticker to user based on multi-turn dialog context and sticker using history of user.

Meaningful Answer Generation of E-Commerce Question-Answering

no code implementations14 Nov 2020 Shen Gao, Xiuying Chen, Zhaochun Ren, Dongyan Zhao, Rui Yan

To generate more meaningful answers, in this paper, we propose a novel generative neural model, called the Meaningful Product Answer Generator (MPAG), which alleviates the safe answer problem by taking product reviews, product attributes, and a prototype answer into consideration.

Answer Generation Question Answering +1

Unsupervised Domain Adaptation of a Pretrained Cross-Lingual Language Model

1 code implementation23 Nov 2020 Juntao Li, Ruidan He, Hai Ye, Hwee Tou Ng, Lidong Bing, Rui Yan

Experimental results show that our proposed method achieves significant performance improvements over the state-of-the-art pretrained cross-lingual language model in the CLCD setting.

Language Modelling Mutual Information Estimation +1

Design and Commissioning of the PandaX-4T Cryogenic Distillation System for Krypton and Radon Removal

no code implementations4 Dec 2020 Xiangyi Cui, Zhou Wang, Yonglin Ju, Xiuli Wang, Huaxuan Liu, Wenbo Ma, Jianglai Liu, Li Zhao, Xiangdong Ji, Shuaijie Li, Rui Yan, Haidong Sha, Peiyao Huang

An online cryogenic distillation system for the removal of krypton and radon from xenon was designed and constructed for PandaX-4T, a highly sensitive dark matter detection experiment.

Instrumentation and Detectors High Energy Physics - Experiment

Interactive Fusion of Multi-level Features for Compositional Activity Recognition

1 code implementation10 Dec 2020 Rui Yan, Lingxi Xie, Xiangbo Shu, Jinhui Tang

To understand a complex action, multiple sources of information, including appearance, positional, and semantic features, need to be integrated.

Action Recognition

Reasoning in Dialog: Improving Response Generation by Context Reading Comprehension

1 code implementation14 Dec 2020 Xiuying Chen, Zhi Cui, Jiayi Zhang, Chen Wei, Jianwei Cui, Bin Wang, Dongyan Zhao, Rui Yan

Hence, in this paper, we propose to improve the response generation performance by examining the model's ability to answer a reading comprehension question, where the question is focused on the omitted information in the dialog.

Multi-Task Learning Reading Comprehension +1

The Style-Content Duality of Attractiveness: Learning to Write Eye-Catching Headlines via Disentanglement

no code implementations14 Dec 2020 Mingzhe Li, Xiuying Chen, Min Yang, Shen Gao, Dongyan Zhao, Rui Yan

In this paper, we propose a Disentanglement-based Attractive Headline Generator (DAHG) that generates headline which captures the attractive content following the attractive style.

Disentanglement

OpenViDial: A Large-Scale, Open-Domain Dialogue Dataset with Visual Contexts

1 code implementation30 Dec 2020 Yuxian Meng, Shuhe Wang, Qinghong Han, Xiaofei Sun, Fei Wu, Rui Yan, Jiwei Li

Based on this dataset, we propose a family of encoder-decoder models leveraging both textual and visual contexts, from coarse-grained image features extracted from CNNs to fine-grained object features extracted from Faster R-CNNs.

Dialogue Generation

Dialogue History Matters! Personalized Response Selectionin Multi-turn Retrieval-based Chatbots

no code implementations17 Mar 2021 Juntao Li, Chang Liu, Chongyang Tao, Zhangming Chan, Dongyan Zhao, Min Zhang, Rui Yan

To fill the gap between these up-to-date methods and the real-world applications, we incorporate user-specific dialogue history into the response selection and propose a personalized hybrid matching network (PHMN).

Representation Learning Retrieval

Modeling Text-visual Mutual Dependency for Multi-modal Dialog Generation

1 code implementation30 May 2021 Shuhe Wang, Yuxian Meng, Xiaofei Sun, Fei Wu, Rongbin Ouyang, Rui Yan, Tianwei Zhang, Jiwei Li

Specifically, we propose to model the mutual dependency between text-visual features, where the model not only needs to learn the probability of generating the next dialog utterance given preceding dialog utterances and visual contexts, but also the probability of predicting the visual features in which a dialog utterance takes place, leading the generated dialog utterance specific to the visual context.

Learning to Organize a Bag of Words into Sentences with Neural Networks: An Empirical Study

no code implementations NAACL 2021 Chongyang Tao, Shen Gao, Juntao Li, Yansong Feng, Dongyan Zhao, Rui Yan

Sequential information, a. k. a., orders, is assumed to be essential for processing a sequence with recurrent neural network or convolutional neural network based encoders.

Sentence

A Pre-training Strategy for Zero-Resource Response Selection in Knowledge-Grounded Conversations

no code implementations ACL 2021 Chongyang Tao, Changyu Chen, Jiazhan Feng, Ji-Rong Wen, Rui Yan

Recently, many studies are emerging towards building a retrieval-based dialogue system that is able to effectively leverage background knowledge (e. g., documents) when conversing with humans.

Language Modelling Retrieval +1

Capturing Relations between Scientific Papers: An Abstractive Model for Related Work Section Generation

1 code implementation ACL 2021 Xiuying Chen, Hind Alamro, Mingzhe Li, Shen Gao, Xiangliang Zhang, Dongyan Zhao, Rui Yan

Hence, in this paper, we propose a Relation-aware Related work Generator (RRG), which generates an abstractive related work from the given multiple scientific papers in the same research area.

Relation

Response Ranking with Multi-types of Deep Interactive Representations in Retrieval-based Dialogues

1 code implementation ACM Transactions on Information Systems 2021 Ruijian Xu, Chongyang Tao, Jiazhan Feng, Wei Wu, Rui Yan, Dongyan Zhao

To tackle these challenges, we propose a representation[K]-interaction[L]-matching framework that explores multiple types of deep interactive representations to build context-response matching models for response selection.

Conversational Response Selection Retrieval

Target-Side Data Augmentation for Sequence Generation

1 code implementation ICLR 2022 Shufang Xie, Ang Lv, Yingce Xia, Lijun Wu, Tao Qin, Rui Yan, Tie-Yan Liu

Autoregressive sequence generation, a prevalent task in machine learning and natural language processing, generates every target token conditioned on both a source input and previously generated target tokens.

Abstractive Text Summarization Data Augmentation +2

TaskDrop: A Competitive Baseline for Continual Learning of Sentiment Classification

no code implementations5 Nov 2021 JianPing Mei, Yilun Zheng, Qianwei Zhou, Rui Yan

In this paper, we study the multi-task sentiment classification problem in the continual learning setting, i. e., a model is sequentially trained to classifier the sentiment of reviews of products in a particular category.

Classification Continual Learning +2

Stylized Dialogue Generation with Multi-Pass Dual Learning

1 code implementation NeurIPS 2021 Jinpeng Li, Yingce Xia, Rui Yan, Hongda Sun, Dongyan Zhao, Tie-Yan Liu

Considering there is no parallel data between the contexts and the responses of target style S1, existing works mainly use back translation to generate stylized synthetic data for training, where the data about context, target style S1 and an intermediate style S0 is used.

Dialogue Generation

Object-aware Video-language Pre-training for Retrieval

1 code implementation CVPR 2022 Alex Jinpeng Wang, Yixiao Ge, Guanyu Cai, Rui Yan, Xudong Lin, Ying Shan, XiaoHu Qie, Mike Zheng Shou

In this work, we present Object-aware Transformers, an object-centric approach that extends video-language transformer to incorporate object representations.

Object Retrieval +2

Video-Text Pre-training with Learned Regions

1 code implementation2 Dec 2021 Rui Yan, Mike Zheng Shou, Yixiao Ge, Alex Jinpeng Wang, Xudong Lin, Guanyu Cai, Jinhui Tang

Video-Text pre-training aims at learning transferable representations from large-scale video-text pairs via aligning the semantics between visual and textual information.

Representation Learning Retrieval +2

Expansion-Squeeze-Excitation Fusion Network for Elderly Activity Recognition

no code implementations21 Dec 2021 Xiangbo Shu, Jiawen Yang, Rui Yan, Yan Song

This work focuses on the task of elderly activity recognition, which is a challenging task due to the existence of individual actions and human-object interactions in elderly activities.

Action Recognition Human-Object Interaction Detection

Strategy Synthesis for Zero-Sum Neuro-Symbolic Concurrent Stochastic Games

no code implementations13 Feb 2022 Rui Yan, Gabriel Santos, Gethin Norman, David Parker, Marta Kwiatkowska

Second, we introduce two novel representations for the value functions and strategies, constant-piecewise-linear (CON-PWL) and constant-piecewise-constant (CON-PWC) respectively, and propose Minimax-action-free PI by extending a recent PI method based on alternating player choices for finite state spaces to Borel state spaces, which does not require normal-form games to be solved.

All in One: Exploring Unified Video-Language Pre-training

1 code implementation CVPR 2023 Alex Jinpeng Wang, Yixiao Ge, Rui Yan, Yuying Ge, Xudong Lin, Guanyu Cai, Jianping Wu, Ying Shan, XiaoHu Qie, Mike Zheng Shou

In this work, we for the first time introduce an end-to-end video-language model, namely \textit{all-in-one Transformer}, that embeds raw video and textual signals into joint representations using a unified backbone architecture.

Ranked #6 on TGIF-Transition on TGIF-QA (using extra training data)

Language Modelling Multiple-choice +10

Revitalize Region Feature for Democratizing Video-Language Pre-training of Retrieval

2 code implementations15 Mar 2022 Guanyu Cai, Yixiao Ge, Binjie Zhang, Alex Jinpeng Wang, Rui Yan, Xudong Lin, Ying Shan, Lianghua He, XiaoHu Qie, Jianping Wu, Mike Zheng Shou

Recent dominant methods for video-language pre-training (VLP) learn transferable representations from the raw pixels in an end-to-end manner to achieve advanced performance on downstream video-language retrieval.

Question Answering Retrieval +4

MISC: A MIxed Strategy-Aware Model Integrating COMET for Emotional Support Conversation

1 code implementation ACL 2022 Quan Tu, Yanran Li, Jianwei Cui, Bin Wang, Ji-Rong Wen, Rui Yan

Applying existing methods to emotional support conversation -- which provides valuable assistance to people who are in need -- has two major limitations: (a) they generally employ a conversation-level emotion label, which is too coarse-grained to capture user's instant mental state; (b) most of them focus on expressing empathy in the response(s) rather than gradually reducing user's distress.

Learning to Express in Knowledge-Grounded Conversation

no code implementations NAACL 2022 Xueliang Zhao, Tingchen Fu, Chongyang Tao, Wei Wu, Dongyan Zhao, Rui Yan

Grounding dialogue generation by extra knowledge has shown great potentials towards building a system capable of replying with knowledgeable and engaging responses.

Dialogue Generation

GNN-encoder: Learning a Dual-encoder Architecture via Graph Neural Networks for Dense Passage Retrieval

no code implementations18 Apr 2022 Jiduan Liu, Jiahao Liu, Yang Yang, Jingang Wang, Wei Wu, Dongyan Zhao, Rui Yan

To enhance the performance of dense retrieval models without loss of efficiency, we propose a GNN-encoder model in which query (passage) information is fused into passage (query) representations via graph neural networks that are constructed by queries and their top retrieved passages.

Natural Questions Passage Retrieval +2

Time Domain Adversarial Voice Conversion for ADD 2022

no code implementations19 Apr 2022 Cheng Wen, Tingwei Guo, Xingjun Tan, Rui Yan, Shuran Zhou, Chuandong Xie, Wei Zou, Xiangang Li

In this paper, we describe our speech generation system for the first Audio Deep Synthesis Detection Challenge (ADD 2022).

Voice Conversion

Audio Deep Fake Detection System with Neural Stitching for ADD 2022

no code implementations19 Apr 2022 Rui Yan, Cheng Wen, Shuran Zhou, Tingwei Guo, Wei Zou, Xiangang Li

This paper describes our best system and methodology for ADD 2022: The First Audio Deep Synthesis Detection Challenge\cite{Yi2022ADD}.

Voice Conversion

OTExtSum: Extractive Text Summarisation with Optimal Transport

1 code implementation Findings (NAACL) 2022 Peggy Tang, Kun Hu, Rui Yan, Lei Zhang, Junbin Gao, Zhiyong Wang

Optimal sentence extraction is conceptualised as obtaining an optimal summary that minimises the transportation cost to a given document regarding their semantic distributions.

Sentence

Towards Lossless ANN-SNN Conversion under Ultra-Low Latency with Dual-Phase Optimization

1 code implementation16 May 2022 ZiMing Wang, Shuang Lian, Yuhao Zhang, Xiaoxin Cui, Rui Yan, Huajin Tang

By evaluating on challenging datasets including CIFAR-10, CIFAR- 100 and ImageNet, the proposed method demonstrates the state-of-the-art performance in terms of accuracy, latency and energy preservation.

object-detection Object Detection +1

Label-Efficient Self-Supervised Federated Learning for Tackling Data Heterogeneity in Medical Imaging

1 code implementation17 May 2022 Rui Yan, Liangqiong Qu, Qingyue Wei, Shih-Cheng Huang, Liyue Shen, Daniel Rubin, Lei Xing, Yuyin Zhou

The collection and curation of large-scale medical datasets from multiple institutions is essential for training accurate deep learning models, but privacy concerns often hinder data sharing.

Federated Learning Privacy Preserving +2

Target-aware Abstractive Related Work Generation with Contrastive Learning

1 code implementation26 May 2022 Xiuying Chen, Hind Alamro, Mingzhe Li, Shen Gao, Rui Yan, Xin Gao, Xiangliang Zhang

The related work section is an important component of a scientific paper, which highlights the contribution of the target paper in the context of the reference papers.

Contrastive Learning TAG

SSM-DTA: Breaking the Barriers of Data Scarcity in Drug-Target Affinity Prediction

2 code implementations20 Jun 2022 Qizhi Pei, Lijun Wu, Jinhua Zhu, Yingce Xia, Shufang Xie, Tao Qin, Haiguang Liu, Tie-Yan Liu, Rui Yan

Accurate prediction of Drug-Target Affinity (DTA) is of vital importance in early-stage drug discovery, facilitating the identification of drugs that can effectively interact with specific targets and regulate their activities.

Drug Discovery Language Modelling +2

RetroGraph: Retrosynthetic Planning with Graph Search

1 code implementation23 Jun 2022 Shufang Xie, Rui Yan, Peng Han, Yingce Xia, Lijun Wu, Chenjuan Guo, Bin Yang, Tao Qin

We observe that the same intermediate molecules are visited many times in the searching process, and they are usually independently treated in previous tree-based methods (e. g., AND-OR tree search, Monte Carlo tree search).

Drug Discovery Multi-step retrosynthesis

Egocentric Video-Language Pretraining @ Ego4D Challenge 2022

1 code implementation4 Jul 2022 Kevin Qinghong Lin, Alex Jinpeng Wang, Mattia Soldan, Michael Wray, Rui Yan, Eric Zhongcong Xu, Difei Gao, RongCheng Tu, Wenzhe Zhao, Weijie Kong, Chengfei Cai, Hongfa Wang, Dima Damen, Bernard Ghanem, Wei Liu, Mike Zheng Shou

In this report, we propose a video-language pretraining (VLP) based solution \cite{kevin2022egovlp} for four Ego4D challenge tasks, including Natural Language Query (NLQ), Moment Query (MQ), Object State Change Classification (OSCC), and PNR Localization (PNR).

Language Modelling Object State Change Classification

Deep Spike Learning with Local Classifiers

1 code implementation IEEE Transactions on Cybernetics 2022 Chenxiang Ma, Rui Yan, Zhaofei Yu, Qiang Yu

We then propose two variants that additionally incorporate temporal dependencies through a backward and forward process, respectively.

Re-creation of Creations: A New Paradigm for Lyric-to-Melody Generation

1 code implementation11 Aug 2022 Ang Lv, Xu Tan, Tao Qin, Tie-Yan Liu, Rui Yan

These characteristics cannot be well handled by neural generation models that learn lyric-to-melody mapping in an end-to-end way, due to several issues: (1) lack of aligned lyric-melody training data to sufficiently learn lyric-melody feature alignment; (2) lack of controllability in generation to better and explicitly align the lyric-melody features.

Language Modelling Retrieval

Towards Efficient Dialogue Pre-training with Transferable and Interpretable Latent Structure

no code implementations22 Oct 2022 Xueliang Zhao, Lemao Liu, Tingchen Fu, Shuming Shi, Dongyan Zhao, Rui Yan

With the availability of massive general-domain dialogue data, pre-trained dialogue generation appears to be super appealing to transfer knowledge from the general domain to downstream applications.

Dialogue Generation

A Low Latency Adaptive Coding Spiking Framework for Deep Reinforcement Learning

no code implementations21 Nov 2022 Lang Qin, Rui Yan, Huajin Tang

In recent years, spiking neural networks (SNNs) have been used in reinforcement learning (RL) due to their low power consumption and event-driven features.

Offline RL reinforcement-learning +1

Neural Machine Translation with Contrastive Translation Memories

1 code implementation6 Dec 2022 Xin Cheng, Shen Gao, Lemao Liu, Dongyan Zhao, Rui Yan

Retrieval-augmented Neural Machine Translation models have been successful in many translation scenarios.

Contrastive Learning Machine Translation +4

Scientific Paper Extractive Summarization Enhanced by Citation Graphs

no code implementations8 Dec 2022 Xiuying Chen, Mingzhe Li, Shen Gao, Rui Yan, Xin Gao, Xiangliang Zhang

We first propose a Multi-granularity Unsupervised Summarization model (MUS) as a simple and low-cost solution to the task.

Extractive Summarization Link Prediction +1

Follow the Timeline! Generating Abstractive and Extractive Timeline Summary in Chronological Order

1 code implementation2 Jan 2023 Xiuying Chen, Mingzhe Li, Shen Gao, Zhangming Chan, Dongyan Zhao, Xin Gao, Xiangliang Zhang, Rui Yan

Nowadays, time-stamped web documents related to a general news query floods spread throughout the Internet, and timeline summarization targets concisely summarizing the evolution trajectory of events along the timeline.

Document Summarization Timeline Summarization +1

EZInterviewer: To Improve Job Interview Performance with Mock Interview Generator

no code implementations3 Jan 2023 Mingzhe Li, Xiuying Chen, Weiheng Liao, Yang song, Tao Zhang, Dongyan Zhao, Rui Yan

The key idea is to reduce the number of parameters that rely on interview dialogs by disentangling the knowledge selector and dialog generator so that most parameters can be trained with ungrounded dialogs as well as the resume data that are not low-resource.

Learning towards Selective Data Augmentation for Dialogue Generation

no code implementations17 Mar 2023 Xiuying Chen, Mingzhe Li, Jiayi Zhang, Xiaoqiang Xia, Chen Wei, Jianwei Cui, Xin Gao, Xiangliang Zhang, Rui Yan

As it is cumbersome and expensive to acquire a huge amount of data for training neural dialog models, data augmentation is proposed to effectively utilize existing training samples.

Data Augmentation Dialogue Generation +1

Attack is Good Augmentation: Towards Skeleton-Contrastive Representation Learning

no code implementations8 Apr 2023 Binqian Xu, Xiangbo Shu, Rui Yan, Guo-Sen Xie, Yixiao Ge, Mike Zheng Shou

In particular, we propose a novel Attack-Augmentation Mixing-Contrastive learning (A$^2$MC) to contrast hard positive features and hard negative features for learning more robust skeleton representations.

Action Recognition Contrastive Learning +4

Graph based Label Enhancement for Multi-instance Multi-label learning

no code implementations21 Apr 2023 Houcheng Su, Jintao Huang, Daixian Liu, Rui Yan, Jiao Li, Chi-Man Vong

Multi-instance multi-label (MIML) learning is widely applicated in numerous domains, such as the image classification where one image contains multiple instances correlated with multiple logic labels simultaneously.

Image Classification Multi-Label Learning

ResiDual: Transformer with Dual Residual Connections

1 code implementation28 Apr 2023 Shufang Xie, Huishuai Zhang, Junliang Guo, Xu Tan, Jiang Bian, Hany Hassan Awadalla, Arul Menezes, Tao Qin, Rui Yan

In this paper, we propose ResiDual, a novel Transformer architecture with Pre-Post-LN (PPLN), which fuses the connections in Post-LN and Pre-LN together and inherits their advantages while avoids their limitations.

Machine Translation

Lift Yourself Up: Retrieval-augmented Text Generation with Self Memory

1 code implementation3 May 2023 Xin Cheng, Di Luo, Xiuying Chen, Lemao Liu, Dongyan Zhao, Rui Yan

In this paper, by exploring the duality of the primal problem: better generation also prompts better memory, we propose a novel framework, selfmem, which addresses this limitation by iteratively employing a retrieval-augmented generator to create an unbounded memory pool and using a memory selector to choose one output as memory for the subsequent generation round.

Abstractive Text Summarization Dialogue Generation +2

GETMusic: Generating Any Music Tracks with a Unified Representation and Diffusion Framework

1 code implementation18 May 2023 Ang Lv, Xu Tan, Peiling Lu, Wei Ye, Shikun Zhang, Jiang Bian, Rui Yan

Our proposed representation, coupled with the non-autoregressive generative model, empowers GETMusic to generate music with any arbitrary source-target track combinations.

Denoising Music Generation

Decouple knowledge from parameters for plug-and-play language modeling

1 code implementation19 May 2023 Xin Cheng, Yankai Lin, Xiuying Chen, Dongyan Zhao, Rui Yan

The key intuition is to decouple the knowledge storage from model parameters with an editable and scalable key-value memory and leverage knowledge in an explainable manner by knowledge retrieval in the DPM.

Domain Adaptation Language Modelling +1

Exploiting Noise as a Resource for Computation and Learning in Spiking Neural Networks

1 code implementation25 May 2023 Gehua Ma, Rui Yan, Huajin Tang

Despite extensive research on spiking neural networks (SNNs), most studies are established on deterministic models, overlooking the inherent non-deterministic, noisy nature of neural computations.

RankCSE: Unsupervised Sentence Representations Learning via Learning to Rank

1 code implementation26 May 2023 Jiduan Liu, Jiahao Liu, Qifan Wang, Jingang Wang, Wei Wu, Yunsen Xian, Dongyan Zhao, Kai Chen, Rui Yan

In this paper, we propose a novel approach, RankCSE, for unsupervised sentence representation learning, which incorporates ranking consistency and ranking distillation with contrastive learning into a unified framework.

Contrastive Learning Learning-To-Rank +4

PreQuant: A Task-agnostic Quantization Approach for Pre-trained Language Models

no code implementations30 May 2023 Zhuocheng Gong, Jiahao Liu, Qifan Wang, Yang Yang, Jingang Wang, Wei Wu, Yunsen Xian, Dongyan Zhao, Rui Yan

While transformer-based pre-trained language models (PLMs) have dominated a number of NLP applications, these models are heavy to deploy and expensive to use.

Quantization

Retrosynthesis Prediction with Local Template Retrieval

no code implementations7 Jun 2023 Shufang Xie, Rui Yan, Junliang Guo, Yingce Xia, Lijun Wu, Tao Qin

Furthermore, we propose a lightweight adapter to adjust the weights when combing neural network and KNN predictions conditioned on the hidden representation and the retrieved templates.

Drug Discovery Retrieval +1

Semi-Offline Reinforcement Learning for Optimized Text Generation

1 code implementation16 Jun 2023 Changyu Chen, Xiting Wang, Yiqiao Jin, Victor Ye Dong, Li Dong, Jie Cao, Yi Liu, Rui Yan

In reinforcement learning (RL), there are two major settings for interacting with the environment: online and offline.

Offline RL reinforcement-learning +2

Temporal Conditioning Spiking Latent Variable Models of the Neural Response to Natural Visual Scenes

no code implementations NeurIPS 2023 Gehua Ma, Runhao Jiang, Rui Yan, Huajin Tang

This work presents the temporal conditioning spiking latent variable models (TeCoS-LVM) to simulate the neural response to natural visual stimuli.

DialoGPS: Dialogue Path Sampling in Continuous Semantic Space for Data Augmentation in Multi-Turn Conversations

no code implementations29 Jun 2023 Ang Lv, Jinpeng Li, Yuhan Chen, Xing Gao, Ji Zhang, Rui Yan

In open-domain dialogue generation tasks, contexts and responses in most datasets are one-to-one mapped, violating an important many-to-many characteristic: a context leads to various responses, and a response answers multiple contexts.

Data Augmentation Dialogue Generation +2

Point-based Value Iteration for Neuro-Symbolic POMDPs

no code implementations30 Jun 2023 Rui Yan, Gabriel Santos, Gethin Norman, David Parker, Marta Kwiatkowska

This requires functions over continuous-state beliefs, for which we propose a novel piecewise linear and convex representation (P-PWLC) in terms of polyhedra covering the continuous-state space and value vectors, and extend Bellman backups to this representation.

Collision Avoidance Decision Making +1

Enhancing Job Recommendation through LLM-based Generative Adversarial Networks

no code implementations20 Jul 2023 Yingpeng Du, Di Luo, Rui Yan, Hongzhi Liu, Yang song, HengShu Zhu, Jie Zhang

However, directly leveraging LLMs to enhance recommendation results is not a one-size-fits-all solution, as LLMs may suffer from fabricated generation and few-shot problems, which degrade the quality of resume completion.

UniVTG: Towards Unified Video-Language Temporal Grounding

1 code implementation ICCV 2023 Kevin Qinghong Lin, Pengchuan Zhang, Joya Chen, Shraman Pramanick, Difei Gao, Alex Jinpeng Wang, Rui Yan, Mike Zheng Shou

Most methods in this direction develop taskspecific models that are trained with type-specific labels, such as moment retrieval (time interval) and highlight detection (worthiness curve), which limits their abilities to generalize to various VTG tasks and labels.

Ranked #3 on Highlight Detection on QVHighlights (using extra training data)

Highlight Detection Moment Retrieval +3

M$^3$Net: Multi-view Encoding, Matching, and Fusion for Few-shot Fine-grained Action Recognition

no code implementations6 Aug 2023 Hao Tang, Jun Liu, Shuanglin Yan, Rui Yan, Zechao Li, Jinhui Tang

Due to the scarcity of manually annotated data required for fine-grained video understanding, few-shot fine-grained (FS-FG) action recognition has gained significant attention, with the aim of classifying novel fine-grained action categories with only a few labeled instances.

Decision Making Fine-grained Action Recognition +1

CharacterChat: Learning towards Conversational AI with Personalized Social Support

1 code implementation20 Aug 2023 Quan Tu, Chuanqi Chen, Jinpeng Li, Yanran Li, Shuo Shang, Dongyan Zhao, Ran Wang, Rui Yan

In our modern, fast-paced, and interconnected world, the importance of mental well-being has grown into a matter of great urgency.

Neuromorphic Auditory Perception by Neural Spiketrum

no code implementations11 Sep 2023 Huajin Tang, Pengjie Gu, Jayawan Wijekoon, MHD Anas Alsakkal, ZiMing Wang, Jiangrong Shen, Rui Yan

Neuromorphic computing holds the promise to achieve the energy efficiency and robust learning performance of biological neural systems.

Dataset Condensation via Generative Model

no code implementations14 Sep 2023 David Junhao Zhang, Heng Wang, Chuhui Xue, Rui Yan, Wenqing Zhang, Song Bai, Mike Zheng Shou

Dataset condensation aims to condense a large dataset with a lot of training samples into a small set.

Dataset Condensation

SCALE: Synergized Collaboration of Asymmetric Language Translation Engines

1 code implementation29 Sep 2023 Xin Cheng, Xun Wang, Tao Ge, Si-Qing Chen, Furu Wei, Dongyan Zhao, Rui Yan

In this paper, we introduce SCALE, a collaborative framework that connects compact Specialized Translation Models (STMs) and general-purpose Large Language Models (LLMs) as one unified translation engine.

Continual Learning Translation

FABind: Fast and Accurate Protein-Ligand Binding

1 code implementation NeurIPS 2023 Qizhi Pei, Kaiyuan Gao, Lijun Wu, Jinhua Zhu, Yingce Xia, Shufang Xie, Tao Qin, Kun He, Tie-Yan Liu, Rui Yan

In this work, we propose $\mathbf{FABind}$, an end-to-end model that combines pocket prediction and docking to achieve accurate and fast protein-ligand binding.

Drug Discovery Pose Estimation +1

Retrieval-based Knowledge Transfer: An Effective Approach for Extreme Large Language Model Compression

no code implementations24 Oct 2023 Jiduan Liu, Jiahao Liu, Qifan Wang, Jingang Wang, Xunliang Cai, Dongyan Zhao, Ran Lucien Wang, Rui Yan

In particular, our approach extracts knowledge from LLMs to construct a knowledge store, from which the small-scale model can retrieve relevant information and leverage it for effective inference.

Language Modelling Large Language Model +3

CycleAlign: Iterative Distillation from Black-box LLM to White-box Models for Better Human Alignment

no code implementations25 Oct 2023 Jixiang Hong, Quan Tu, Changyu Chen, Xing Gao, Ji Zhang, Rui Yan

With in-context learning (ICL) as the core of the cycle, the black-box models are able to rank the model-generated responses guided by human-craft instruction and demonstrations about their preferences.

In-Context Learning Instruction Following +2

From Indeterminacy to Determinacy: Augmenting Logical Reasoning Capabilities with Large Language Models

1 code implementation28 Oct 2023 Hongda Sun, Weikai Xu, Wei Liu, Jian Luan, Bin Wang, Shuo Shang, Ji-Rong Wen, Rui Yan

To address these challenges, we propose DetermLR, a novel reasoning framework that formulates the reasoning process as a transformational journey from indeterminate premises to determinate ones.

Logical Reasoning

Improving Input-label Mapping with Demonstration Replay for In-context Learning

no code implementations30 Oct 2023 Zhuocheng Gong, Jiahao Liu, Qifan Wang, Jingang Wang, Xunliang Cai, Dongyan Zhao, Rui Yan

The effectiveness of ICL can be attributed to the strong language modeling capabilities of large language models (LLMs), which enable them to learn the mapping between input and labels based on in-context demonstrations.

In-Context Learning Language Modelling

Multiplayer Homicidal Chauffeur Reach-Avoid Games: A Pursuit Enclosure Function Approach

no code implementations4 Nov 2023 Rui Yan, Xiaoming Duan, Rui Zou, Xin He, Zongying Shi, Francesco Bullo

We propose a cooperative strategy for the pursuers based on subgames for multiple pursuers against one evader and optimal task allocation.

ERP

Are We Falling in a Middle-Intelligence Trap? An Analysis and Mitigation of the Reversal Curse

1 code implementation13 Nov 2023 Ang Lv, Kaiyi Zhang, Shufang Xie, Quan Tu, Yuhan Chen, Ji-Rong Wen, Rui Yan

Recent studies have highlighted a phenomenon in large language models (LLMs) known as "the reversal curse," in which the order of knowledge entities in the training data biases the models' comprehension.

Denoising Language Modelling

Fortify the Shortest Stave in Attention: Enhancing Context Awareness of Large Language Models for Effective Tool Use

1 code implementation7 Dec 2023 Yuhan Chen, Ang Lv, Ting-En Lin, Changyu Chen, Yuchuan Wu, Fei Huang, Yongbin Li, Rui Yan

Specifically, the crucial information in the context will be potentially overlooked by model when it is positioned in the trough zone of the attention waveform, leading to decreased performance.

Trajectory Planning

Collaborative Synthesis of Patient Records through Multi-Visit Health State Inference

1 code implementation22 Dec 2023 Hongda Sun, Hongzhan Lin, Rui Yan

Furthermore, we propose to generate medical reports to add textual descriptions for each medical event, providing broader applications for synthesized EHR data.

Common Sense Reasoning

CharacterEval: A Chinese Benchmark for Role-Playing Conversational Agent Evaluation

1 code implementation2 Jan 2024 Quan Tu, Shilong Fan, Zihang Tian, Rui Yan

Recently, the advent of large language models (LLMs) has revolutionized generative agents.

Batch-ICL: Effective, Efficient, and Order-Agnostic In-Context Learning

1 code implementation12 Jan 2024 Kaiyi Zhang, Ang Lv, Yuhan Chen, Hansen Ha, Tao Xu, Rui Yan

In this paper, by treating in-context learning (ICL) as a meta-optimization process, we explain why LLMs are sensitive to the order of ICL examples.

In-Context Learning Zero-Shot Learning

BioT5+: Towards Generalized Biological Understanding with IUPAC Integration and Multi-task Tuning

1 code implementation27 Feb 2024 Qizhi Pei, Lijun Wu, Kaiyuan Gao, Xiaozhuan Liang, Yin Fang, Jinhua Zhu, Shufang Xie, Tao Qin, Rui Yan

However, previous efforts like BioT5 faced challenges in generalizing across diverse tasks and lacked a nuanced understanding of molecular structures, particularly in their textual representations (e. g., IUPAC).

Molecule Captioning Text-based de novo Molecule Generation

Leveraging Biomolecule and Natural Language through Multi-Modal Learning: A Survey

2 code implementations3 Mar 2024 Qizhi Pei, Lijun Wu, Kaiyuan Gao, Jinhua Zhu, Yue Wang, Zun Wang, Tao Qin, Rui Yan

The integration of biomolecular modeling with natural language (BL) has emerged as a promising interdisciplinary area at the intersection of artificial intelligence, chemistry and biology.

Property Prediction

Masked Thought: Simply Masking Partial Reasoning Steps Can Improve Mathematical Reasoning Learning of Language Models

1 code implementation4 Mar 2024 Changyu Chen, Xiting Wang, Ting-En Lin, Ang Lv, Yuchuan Wu, Xin Gao, Ji-Rong Wen, Rui Yan, Yongbin Li

In reasoning tasks, even a minor error can cascade into inaccurate results, leading to suboptimal performance of large language models in such domains.

Data Augmentation GSM8K +2

"In Dialogues We Learn": Towards Personalized Dialogue Without Pre-defined Profiles through In-Dialogue Learning

no code implementations5 Mar 2024 Chuanqi Cheng, Quan Tu, Wei Wu, Shuo Shang, Cunli Mao, Zhengtao Yu, Rui Yan

Personalized dialogue systems have gained significant attention in recent years for their ability to generate responses in alignment with different personas.

Dialogue Generation

Pursuit Winning Strategies for Reach-Avoid Games with Polygonal Obstacles

no code implementations10 Mar 2024 Rui Yan, Shuai Mi, Xiaoming Duan, Jintao Chen, Xiangyang Ji

The pursuers cooperate to protect a convex region from the evaders who try to reach the region.

What Makes Quantization for Large Language Models Hard? An Empirical Study from the Lens of Perturbation

no code implementations11 Mar 2024 Zhuocheng Gong, Jiahao Liu, Jingang Wang, Xunliang Cai, Dongyan Zhao, Rui Yan

Our findings reveal several connections between the properties of perturbations and LLM performance, providing insights into the failure cases of uniform quantization and suggesting potential solutions to improve the robustness of LLM quantization.

Computational Efficiency Quantization

StreamingDialogue: Prolonged Dialogue Learning via Long Context Compression with Minimal Losses

no code implementations13 Mar 2024 Jia-Nan Li, Quan Tu, Cunli Mao, Zhengtao Yu, Ji-Rong Wen, Rui Yan

Accordingly, we introduce StreamingDialogue, which compresses long dialogue history into conv-attn sinks with minimal losses, and thus reduces computational complexity quadratically with the number of sinks (i. e., the number of utterances).

From Skepticism to Acceptance: Simulating the Attitude Dynamics Toward Fake News

no code implementations14 Mar 2024 YuHan Liu, Xiuying Chen, Xiaoqing Zhang, Xing Gao, Ji Zhang, Rui Yan

Our simulation results uncover patterns in fake news propagation related to topic relevance, and individual traits, aligning with real-world observations.

StyleChat: Learning Recitation-Augmented Memory in LLMs for Stylized Dialogue Generation

no code implementations18 Mar 2024 Jinpeng Li, Zekai Zhang, Quan Tu, Xin Cheng, Dongyan Zhao, Rui Yan

Furthermore, although many prompt-based methods have been proposed to accomplish specific tasks, their performance in complex real-world scenarios involving a wide variety of dialog styles further enhancement.

Dialogue Generation

Selecting Query-bag as Pseudo Relevance Feedback for Information-seeking Conversations

no code implementations22 Mar 2024 Xiaoqing Zhang, Xiuying Chen, Shen Gao, Shuqi Li, Xin Gao, Ji-Rong Wen, Rui Yan

Given the user query, the information-seeking dialogue systems first retrieve a subset of response candidates, then further select the best response from the candidate set through re-ranking.

Contrastive Learning Re-Ranking

Interpreting Key Mechanisms of Factual Recall in Transformer-Based Language Models

1 code implementation28 Mar 2024 Ang Lv, Kaiyi Zhang, Yuhan Chen, Yulong Wang, Lifeng Liu, Ji-Rong Wen, Jian Xie, Rui Yan

In this paper, we deeply explore the mechanisms employed by Transformer-based language models in factual recall tasks.

There Are a Thousand Hamlets in a Thousand People’s Eyes: Enhancing Knowledge-grounded Dialogue with Personal Memory

no code implementations ACL 2022 Tingchen Fu, Xueliang Zhao, Chongyang Tao, Ji-Rong Wen, Rui Yan

Knowledge-grounded conversation (KGC) shows great potential in building an engaging and knowledgeable chatbot, and knowledge selection is a key ingredient in it.

Chatbot

ProphetChat: Enhancing Dialogue Generation with Simulation of Future Conversation

no code implementations ACL 2022 Chang Liu, Xu Tan, Chongyang Tao, Zhenxin Fu, Dongyan Zhao, Tie-Yan Liu, Rui Yan

To enable the chatbot to foresee the dialogue future, we design a beam-search-like roll-out strategy for dialogue future simulation using a typical dialogue generation model and a dialogue selector.

Dialogue Generation Response Generation

Plan-CVAE: A Planning-based Conditional Variational Autoencoder for Story Generation

no code implementations CCL 2020 Lin Wang, Juntao Li, Rui Yan, Dongyan Zhao

Story generation is a challenging task of automatically creating natural languages to describe a sequence of events, which requires outputting text with not only a consistent topic but also novel wordings.

Story Generation

Finding the Dominant Winning Ticket in Pre-Trained Language Models

no code implementations Findings (ACL) 2022 Zhuocheng Gong, Di He, Yelong Shen, Tie-Yan Liu, Weizhu Chen, Dongyan Zhao, Ji-Rong Wen, Rui Yan

Empirically, we show that (a) the dominant winning ticket can achieve performance that is comparable with that of the full-parameter model, (b) the dominant winning ticket is transferable across different tasks, (c) and the dominant winning ticket has a natural structure within each parameter matrix.

Amalgamating Knowledge from Two Teachers for Task-oriented Dialogue System with Adversarial Training

1 code implementation EMNLP 2020 Wanwei He, Min Yang, Rui Yan, Chengming Li, Ying Shen, Ruifeng Xu

Instead of adopting the classic student-teacher learning of forcing the output of a student network to exactly mimic the soft targets produced by the teacher networks, we introduce two discriminators as in generative adversarial network (GAN) to transfer knowledge from two teachers to the student.

Generative Adversarial Network Task-Oriented Dialogue Systems

Unsupervised Mitigating Gender Bias by Character Components: A Case Study of Chinese Word Embedding

no code implementations NAACL (GeBNLP) 2022 Xiuying Chen, Mingzhe Li, Rui Yan, Xin Gao, Xiangliang Zhang

Word embeddings learned from massive text collections have demonstrated significant levels of discriminative biases. However, debias on the Chinese language, one of the most spoken languages, has been less explored. Meanwhile, existing literature relies on manually created supplementary data, which is time- and energy-consuming. In this work, we propose the first Chinese Gender-neutral word Embedding model (CGE) based on Word2vec, which learns gender-neutral word embeddings without any labeled data. Concretely, CGE utilizes and emphasizes the rich feminine and masculine information contained in radicals, i. e., a kind of component in Chinese characters, during the training procedure. This consequently alleviates discriminative gender biases. Experimental results on public benchmark datasets show that our unsupervised method outperforms the state-of-the-art supervised debiased word embedding models without sacrificing the functionality of the embedding model.

Word Embeddings

Cannot find the paper you are looking for? You can Submit a new open access paper.