Search Results for author: Yang Feng

Found 82 papers, 19 papers with code

Universal Simultaneous Machine Translation with Mixture-of-Experts Wait-k Policy

no code implementations11 Sep 2021 Shaolei Zhang, Yang Feng

Simultaneous machine translation (SiMT) generates translation before reading the entire source sentence and hence it has to trade off between translation quality and latency.

Machine Translation Translation

Modeling Concentrated Cross-Attention for Neural Machine Translation with Gaussian Mixture Model

no code implementations11 Sep 2021 Shaolei Zhang, Yang Feng

Cross-attention is an important component of neural machine translation (NMT), which is always realized by dot-product attention in previous methods.

Machine Translation Translation

Mixup Decoding for Diverse Machine Translation

no code implementations8 Sep 2021 Jicheng Li, Pengzhi Gao, Xuanfu Wu, Yang Feng, Zhongjun He, Hua Wu, Haifeng Wang

To further improve the faithfulness and diversity of the translations, we propose two simple but effective approaches to select diverse sentence pairs in the training corpus and adjust the interpolation weight for each pair correspondingly.

Machine Translation Translation

Towards Expressive Communication with Internet Memes: A New Multimodal Conversation Dataset and Benchmark

1 code implementation4 Sep 2021 Zhengcong Fei, Zekang Li, Jinchao Zhang, Yang Feng, Jie zhou

Compared to previous dialogue tasks, MOD is much more challenging since it requires the model to understand the multimodal elements as well as the emotions behind them.

Importance-based Neuron Allocation for Multilingual Neural Machine Translation

1 code implementation ACL 2021 Wanying Xie, Yang Feng, Shuhao Gu, Dong Yu

Multilingual neural machine translation with a single model has drawn much attention due to its capability to deal with multiple languages.

Machine Translation Translation

Sequence-Level Training for Non-Autoregressive Neural Machine Translation

1 code implementation15 Jun 2021 Chenze Shao, Yang Feng, Jinchao Zhang, Fandong Meng, Jie zhou

Non-Autoregressive Neural Machine Translation (NAT) removes the autoregressive mechanism and achieves significant decoding speedup through generating target words independently and simultaneously.

Machine Translation Translation

GTM: A Generative Triple-Wise Model for Conversational Question Generation

no code implementations ACL 2021 Lei Shen, Fandong Meng, Jinchao Zhang, Yang Feng, Jie zhou

Generating some appealing questions in open-domain conversations is an effective way to improve human-machine interactions and lead the topic to a broader or deeper direction.

Question Generation

Conversations Are Not Flat: Modeling the Dynamic Information Flow across Dialogue Utterances

1 code implementation ACL 2021 Zekang Li, Jinchao Zhang, Zhengcong Fei, Yang Feng, Jie zhou

Nowadays, open-domain dialogue models can generate acceptable responses according to the historical context based on the large-scale pre-trained language models.

Dialogue Generation

Addressing Inquiries about History: An Efficient and Practical Framework for Evaluating Open-domain Chatbot Consistency

1 code implementation4 Jun 2021 Zekang Li, Jinchao Zhang, Zhengcong Fei, Yang Feng, Jie zhou

Employing human judges to interact with chatbots on purpose to check their capacities is costly and low-efficient, and difficult to get rid of subjective bias.

Chatbot Natural Language Inference

Transfer Learning under High-dimensional Generalized Linear Models

no code implementations29 May 2021 Ye Tian, Yang Feng

When we don't know which sources to transfer, an algorithm-free transferable source detection approach is introduced to detect informative sources.

Transfer Learning

Machine Collaboration

no code implementations6 May 2021 Qingfeng Liu, Yang Feng

We propose a new ensemble framework for supervised learning, called machine collaboration (MaC), using a collection of base machines for prediction tasks.

SE-DAE: Style-Enhanced Denoising Auto-Encoder for Unsupervised Text Style Transfer

no code implementations27 Apr 2021 Jicheng Li, Yang Feng, Jiao Ou

Moreover, to alleviate the conflict between the targets of the conventional denoising procedure and the style transfer task, we propose another novel style denoising mechanism, which is more compatible with the target of the style transfer task.

Denoising Style Transfer +2

Modeling Coverage for Non-Autoregressive Neural Machine Translation

no code implementations24 Apr 2021 Yong Shan, Yang Feng, Chenze Shao

Non-Autoregressive Neural Machine Translation (NAT) has achieved significant inference speedup by generating all tokens simultaneously.

Machine Translation Translation

Pruning-then-Expanding Model for Domain Adaptation of Neural Machine Translation

1 code implementation NAACL 2021 Shuhao Gu, Yang Feng, Wanying Xie

Domain Adaptation is widely used in practical applications of neural machine translation, which aims to achieve good performance on both the general-domain and in-domain.

Domain Adaptation Knowledge Distillation +2

Learning to Select Context in a Hierarchical and Global Perspective for Open-domain Dialogue Generation

no code implementations18 Feb 2021 Lei Shen, Haolan Zhan, Xin Shen, Yang Feng

Open-domain multi-turn conversations mainly have three features, which are hierarchical semantic structure, redundant information, and long-term dependency.

Dialogue Generation

RaSE: A Variable Screening Framework via Random Subspace Ensembles

1 code implementation7 Feb 2021 Ye Tian, Yang Feng

Variable screening methods have been shown to be effective in dimension reduction under the ultra-high dimensional setting.

Dimensionality Reduction

WeChat AI & ICT's Submission for DSTC9 Interactive Dialogue Evaluation Track

no code implementations20 Jan 2021 Zekang Li, Zongjia Li, Jinchao Zhang, Yang Feng, Jie zhou

We participate in the DSTC9 Interactive Dialogue Evaluation Track (Gunasekara et al. 2020) sub-task 1 (Knowledge Grounded Dialogue) and sub-task 2 (Interactive Dialogue).

Dialogue Evaluation Language Modelling

The Interplay of Demographic Variables and Social Distancing Scores in Deep Prediction of U.S. COVID-19 Cases

no code implementations6 Jan 2021 Francesca Tang, Yang Feng, Hamza Chiheb, Jianqing Fan

With the severity of the COVID-19 outbreak, we characterize the nature of the growth trajectories of counties in the United States using a novel combination of spectral clustering and the correlation matrix.

Knowledge Distillation based Ensemble Learning for Neural Machine Translation

no code implementations1 Jan 2021 Chenze Shao, Meng Sun, Yang Feng, Zhongjun He, Hua Wu, Haifeng Wang

Under this framework, we introduce word-level ensemble learning and sequence-level ensemble learning for neural machine translation, where sequence-level ensemble learning is capable of aggregating translation models with different decoding strategies.

Ensemble Learning Knowledge Distillation +2

Future-Guided Incremental Transformer for Simultaneous Translation

no code implementations23 Dec 2020 Shaolei Zhang, Yang Feng, Liangyou Li

Simultaneous translation (ST) starts translations synchronously while reading source sentences, and is used in many online scenarios.

Knowledge Distillation Translation

Spectral clustering via adaptive layer aggregation for multi-layer networks

no code implementations7 Dec 2020 Sihan Huang, Haolei Weng, Yang Feng

One of the fundamental problems in network analysis is detecting community structure in multi-layer networks, of which each layer represents one type of edge information among the nodes.

Community Detection

Investigating Catastrophic Forgetting During Continual Training for Neural Machine Translation

no code implementations COLING 2020 Shuhao Gu, Yang Feng

The investigation on the modules of the NMT model shows that some modules have tight relation with the general-domain knowledge while some other modules are more essential in the domain adaptation.

Domain Adaptation Machine Translation +1

Generating Diverse Translation from Model Distribution with Dropout

no code implementations EMNLP 2020 Xuanfu Wu, Yang Feng, Chenze Shao

Despite the improvement of translation quality, neural machine translation (NMT) often suffers from the lack of diversity in its generation.

Machine Translation Translation +1

Token-level Adaptive Training for Neural Machine Translation

1 code implementation EMNLP 2020 Shuhao Gu, Jinchao Zhang, Fandong Meng, Yang Feng, Wanying Xie, Jie zhou, Dong Yu

The vanilla NMT model usually adopts trivial equal-weighted objectives for target tokens with different frequencies and tends to generate more high-frequency tokens and less low-frequency tokens compared with the golden token distribution.

Machine Translation Translation

Universal Model for Multi-Domain Medical Image Retrieval

no code implementations14 Jul 2020 Yang Feng, Yubao Liu, Jiebo Luo

Usually, one image retrieval model is only trained to handle images from one modality or one source.

Medical Image Retrieval

Robust Neural Machine Translation with ASR Errors

no code implementations WS 2020 Haiyang Xue, Yang Feng, Shuhao Gu, Wei Chen

In this paper, we propose a method to handle the two problems so as to generate robust translation to ASR errors.

automatic-speech-recognition Machine Translation +2

RaSE: Random Subspace Ensemble Classification

no code implementations16 Jun 2020 Ye Tian, Yang Feng

In addition, we show that in a high-dimensional framework, the number of random subspaces needs to be very large to guarantee that a subspace covering signals is selected.

Classification General Classification

Nested Model Averaging on Solution Path for High-dimensional Linear Regression

no code implementations16 May 2020 Yang Feng, Qing-Feng Liu

We study the nested model averaging method on the solution path for a high-dimensional linear regression problem.

CDL: Curriculum Dual Learning for Emotion-Controllable Response Generation

no code implementations ACL 2020 Lei Shen, Yang Feng

Emotion-controllable response generation is an attractive and valuable task that aims to make open-domain conversations more empathetic and engaging.

Curriculum Learning

Towards Multimodal Response Generation with Exemplar Augmentation and Curriculum Optimization

no code implementations26 Apr 2020 Zeyang Lei, Zekang Li, Jinchao Zhang, Fandong Meng, Yang Feng, Yujiu Yang, Cheng Niu, Jie zhou

Furthermore, to facilitate the convergence of Gaussian mixture prior and posterior distributions, we devise a curriculum optimization strategy to progressively train the model under multiple training criteria from easy to hard.

Imbalanced classification: a paradigm-based review

no code implementations11 Feb 2020 Yang Feng, Min Zhou, Xin Tong

For each pair of resampling techniques and classification methods, we use simulation studies and a real data set on credit card fraud to study the performance under different evaluation metrics.

Classification General Classification +1

Bridging Text and Video: A Universal Multimodal Transformer for Video-Audio Scene-Aware Dialog

1 code implementation1 Feb 2020 Zekang Li, Zongjia Li, Jinchao Zhang, Yang Feng, Cheng Niu, Jie zhou

Audio-Visual Scene-Aware Dialog (AVSD) is a task to generate responses when chatting about a given video, which is organized as a track of the 8th Dialog System Technology Challenge (DSTC8).

Dialogue Generation Multi-Task Learning

Modeling Fluency and Faithfulness for Diverse Neural Machine Translation

1 code implementation30 Nov 2019 Yang Feng, Wanying Xie, Shuhao Gu, Chenze Shao, Wen Zhang, Zhengxin Yang, Dong Yu

Neural machine translation models usually adopt the teacher forcing strategy for training which requires the predicted sequence matches ground truth word by word and forces the probability of each prediction to approach a 0-1 distribution.

Machine Translation Translation

Minimizing the Bag-of-Ngrams Difference for Non-Autoregressive Neural Machine Translation

1 code implementation21 Nov 2019 Chenze Shao, Jinchao Zhang, Yang Feng, Fandong Meng, Jie zhou

Non-Autoregressive Neural Machine Translation (NAT) achieves significant decoding speedup through generating target words independently and simultaneously.

Machine Translation Translation

Improving Bidirectional Decoding with Dynamic Target Semantics in Neural Machine Translation

no code implementations5 Nov 2019 Yong Shan, Yang Feng, Jinchao Zhang, Fandong Meng, Wen Zhang

Generally, Neural Machine Translation models generate target words in a left-to-right (L2R) manner and fail to exploit any future (right) semantics information, which usually produces an unbalanced translation.

Machine Translation Translation

Software Engineering Practice in the Development of Deep Learning Applications

no code implementations8 Oct 2019 Xufan Zhang, Yilin Yang, Yang Feng, Zhenyu Chen

Specifically, we asked the respondents to identify lacks and challenges in the practice of the development life cycle of DL applications.

Software Engineering

Improving Multi-Head Attention with Capsule Networks

no code implementations31 Aug 2019 Shuhao Gu, Yang Feng

Multi-head attention advances neural machine translation by working out multiple versions of attention in different subspaces, but the neglect of semantic overlapping between subspaces increases the difficulty of translation and consequently hinders the further improvement of translation performance.

Machine Translation Translation

Incremental Transformer with Deliberation Decoder for Document Grounded Conversations

1 code implementation ACL 2019 Zekang Li, Cheng Niu, Fandong Meng, Yang Feng, Qian Li, Jie zhou

Document Grounded Conversations is a task to generate dialogue responses when chatting about the content of a given document.

Retrieving Sequential Information for Non-Autoregressive Neural Machine Translation

3 code implementations ACL 2019 Chenze Shao, Yang Feng, Jinchao Zhang, Fandong Meng, Xilin Chen, Jie zhou

Non-Autoregressive Transformer (NAT) aims to accelerate the Transformer model through discarding the autoregressive mechanism and generating target words independently, which fails to exploit the target sequential information.

Machine Translation Translation

Modeling Semantic Relationship in Multi-turn Conversations with Hierarchical Latent Variables

no code implementations ACL 2019 Lei Shen, Yang Feng, Haolan Zhan

Multi-turn conversations consist of complex semantic structures, and it is still a challenge to generate coherent and diverse responses given previous utterances.

Bridging the Gap between Training and Inference for Neural Machine Translation

no code implementations ACL 2019 Wen Zhang, Yang Feng, Fandong Meng, Di You, Qun Liu

Neural Machine Translation (NMT) generates target words sequentially in the way of predicting the next word conditioned on the context words.

Machine Translation Translation

Spatio-temporal Video Re-localization by Warp LSTM

no code implementations CVPR 2019 Yang Feng, Lin Ma, Wei Liu, Jiebo Luo

The need for efficiently finding the video content a user wants is increasing because of the erupting of user-generated videos on the Web.

Video Retrieval

Improving Domain Adaptation Translation with Domain Invariant and Specific Information

no code implementations NAACL 2019 Shuhao Gu, Yang Feng, Qun Liu

Besides, we add a discriminator to the shared encoder and employ adversarial training for the whole model to reinforce the performance of information separation and machine translation simultaneously.

Domain Adaptation Machine Translation +1

Unsupervised Image Captioning

1 code implementation CVPR 2019 Yang Feng, Lin Ma, Wei Liu, Jiebo Luo

Instead of relying on manually labeled image-sentence pairs, our proposed model merely requires an image set, a sentence corpus, and an existing visual concept detector.

Image Captioning

Improving the Robustness of Speech Translation

no code implementations2 Nov 2018 Xiang Li, Haiyang Xue, Wei Chen, Yang Liu, Yang Feng, Qun Liu

Although neural machine translation (NMT) has achieved impressive progress recently, it is usually trained on the clean parallel data set and hence cannot work well when the input sentence is the production of the automatic speech recognition (ASR) system due to the enormous errors in the source.

automatic-speech-recognition Machine Translation +2

Greedy Search with Probabilistic N-gram Matching for Neural Machine Translation

1 code implementation EMNLP 2018 Chenze Shao, Yang Feng, Xilin Chen

Neural machine translation (NMT) models are usually trained with the word-level loss using the teacher forcing algorithm, which not only evaluates the translation improperly but also suffers from exposure bias.

Machine Translation Translation

Video Re-localization

1 code implementation ECCV 2018 Yang Feng, Lin Ma, Wei Liu, Tong Zhang, Jiebo Luo

We first exploit and reorganize the videos in ActivityNet to form a new dataset for video re-localization research, which consists of about 10, 000 videos of diverse visual appearances associated with localized boundary information.

Copy Detection

A likelihood-ratio type test for stochastic block models with bounded degrees

no code implementations12 Jul 2018 Mingao Yuan, Yang Feng, Zuofeng Shang

A fundamental problem in network data analysis is to test Erd\"{o}s-R\'{e}nyi model $\mathcal{G}\left(n,\frac{a+b}{2n}\right)$ versus a bisection stochastic block model $\mathcal{G}\left(n,\frac{a}{n},\frac{b}{n}\right)$, where $a, b>0$ are constants that represent the expected degrees of the graphs and $n$ denotes the number of nodes.

Community Detection Stochastic Block Model

Pairwise Covariates-adjusted Block Model for Community Detection

no code implementations10 Jul 2018 Sihan Huang, Yang Feng

It is shown that both the coefficient estimates of the covariates and the community assignments are consistent under suitable sparsity conditions.

Community Detection Stochastic Block Model

Knowledge Diffusion for Neural Dialogue Generation

1 code implementation ACL 2018 Shuman Liu, Hongshen Chen, Zhaochun Ren, Yang Feng, Qun Liu, Dawei Yin

Our empirical study on a real-world dataset prove that our model is capable of generating meaningful, diverse and natural responses for both factoid-questions and knowledge grounded chi-chats.

Dialogue Generation Question Answering +1

Refining Source Representations with Relation Networks for Neural Machine Translation

no code implementations COLING 2018 Wen Zhang, Jiawei Hu, Yang Feng, Qun Liu

Although neural machine translation with the encoder-decoder framework has achieved great success recently, it still suffers drawbacks of forgetting distant information, which is an inherent disadvantage of recurrent neural network structure, and disregarding relationship between source words during encoding step.

Machine Translation Translation

Large-Scale Model Selection with Misspecification

no code implementations17 Mar 2018 Emre Demirkaya, Yang Feng, Pallavi Basu, Jinchi Lv

Our new information criterion characterizes the impacts of both model misspecification and high dimensionality on model selection.

Model Selection

Unsupervised Anomaly Detection via Variational Auto-Encoder for Seasonal KPIs in Web Applications

7 code implementations12 Feb 2018 Haowen Xu, Wenxiao Chen, Nengwen Zhao, Zeyan Li, Jiahao Bu, Zhihan Li, Ying Liu, Youjian Zhao, Dan Pei, Yang Feng, Jie Chen, Zhaogang Wang, Honglin Qiao

To ensure undisrupted business, large Internet companies need to closely monitor various KPIs (e. g., Page Views, number of online users, and number of orders) of its Web applications, to accurately detect anomalies and trigger timely troubleshooting/mitigation.

Unsupervised Anomaly Detection

Neyman-Pearson classification: parametrics and sample size requirement

no code implementations7 Feb 2018 Xin Tong, Lucy Xia, Jiacheng Wang, Yang Feng

In this work, we employ the parametric linear discriminant analysis (LDA) model and propose a new parametric thresholding algorithm, which does not need the minimum sample size requirements on class $0$ observations and thus is suitable for small sample applications such as rare disease diagnosis.

Classification General Classification

On the estimation of correlation in a binary sequence model

no code implementations27 Dec 2017 Haolei Weng, Yang Feng

We consider a binary sequence generated by thresholding a hidden continuous sequence.

Nonparametric Independence Screening via Favored Smoothing Bandwidth

no code implementations28 Nov 2017 Yang Feng, Yi-Chao Wu, Leonard Stefanski

As a first step, we propose a fast screening method based on the favored smoothing bandwidth of the marginal local constant regression.

Model Selection

Refining Source Representations with Relation Networks for Neural Machine Translation

no code implementations12 Sep 2017 Wen Zhang, Jiawei Hu, Yang Feng, Qun Liu

Although neural machine translation (NMT) with the encoder-decoder framework has achieved great success in recent times, it still suffers from some drawbacks: RNNs tend to forget old information which is often useful and the encoder only operates through words without considering word relationship.

Machine Translation Translation

Information-Propogation-Enhanced Neural Machine Translation by Relation Model

no code implementations6 Sep 2017 Wen Zhang, Jiawei Hu, Yang Feng, Qun Liu

Even though sequence-to-sequence neural machine translation (NMT) model have achieved state-of-art performance in the recent fewer years, but it is widely concerned that the recurrent neural network (RNN) units are very hard to capture the long-distance state information, which means RNN can hardly find the feature with long term dependency as the sequence becomes longer.

Machine Translation Translation

Memory-augmented Neural Machine Translation

no code implementations EMNLP 2017 Yang Feng, Shiyue Zhang, Andi Zhang, Dong Wang, Andrew Abel

Neural machine translation (NMT) has achieved notable success in recent times, however it is also widely recognized that this approach has limitations with handling infrequent words and word pairs.

Machine Translation Translation

Flexible and Creative Chinese Poetry Generation Using Neural Memory

no code implementations ACL 2017 Jiyuan Zhang, Yang Feng, Dong Wang, Yang Wang, Andrew Abel, Shiyue Zhang, Andi Zhang

It has been shown that Chinese poems can be successfully generated by sequence-to-sequence neural models, particularly with the attention mechanism.

Do They All Look the Same? Deciphering Chinese, Japanese and Koreans by Fine-Grained Deep Learning

no code implementations6 Oct 2016 Yu Wang, Haofu Liao, Yang Feng, Xiangyang Xu, Jiebo Luo

We find that Chinese, Japanese and Koreans do exhibit substantial differences in certain attributes, such as bangs, smiling, and bushy eyebrows.

Memory Visualization for Gated Recurrent Neural Networks in Speech Recognition

no code implementations28 Sep 2016 Zhiyuan Tang, Ying Shi, Dong Wang, Yang Feng, Shiyue Zhang

Recurrent neural networks (RNNs) have shown clear superiority in sequence modeling, particularly the ones with gated units, such as long short-term memory (LSTM) and gated recurrent unit (GRU).

automatic-speech-recognition Speech Recognition

Collaborative Learning for Language and Speaker Recognition

no code implementations27 Sep 2016 Lantian Li, Zhiyuan Tang, Dong Wang, Andrew Abel, Yang Feng, Shiyue Zhang

This paper presents a unified model to perform language and speaker recognition simultaneously and altogether.

Speaker Recognition

When Do Luxury Cars Hit the Road? Findings by A Big Data Approach

no code implementations10 May 2016 Yang Feng, Jiebo Luo

Based on the recognition results, we present a data-driven analysis on the relationship between car makes and their appearing times, with implications on lifestyles.

Neyman-Pearson Classification under High-Dimensional Settings

no code implementations13 Aug 2015 Anqi Zhao, Yang Feng, Lie Wang, Xin Tong

Most existing binary classification methods target on the optimization of the overall classification risk and may fail to serve some real-world applications such as cancer diagnosis, where users are more concerned with the risk of misclassifying one specific class than the other.

Classification General Classification

A Projection Based Conditional Dependence Measure with Applications to High-dimensional Undirected Graphical Models

no code implementations7 Jan 2015 Jianqing Fan, Yang Feng, Lucy Xia

Measuring conditional dependence is an important topic in statistics with broad applications including graphical models.

Model Selection in High-Dimensional Misspecified Models

no code implementations23 Dec 2014 Pallavi Basu, Yang Feng, Jinchi Lv

Model selection is indispensable to high-dimensional sparse modeling in selecting the best set of covariates among a sequence of candidate models.

Model Selection

How Many Communities Are There?

no code implementations4 Dec 2014 Diego Franco Saldana, Yi Yu, Yang Feng

Stochastic blockmodels and variants thereof are among the most widely used approaches to community detection for social networks and relational data.

Community Detection Model Selection

Feature Augmentation via Nonparametrics and Selection (FANS) in High Dimensional Classification

no code implementations31 Dec 2013 Jianqing Fan, Yang Feng, Jiancheng Jiang, Xin Tong

We motivate FANS by generalizing the Naive Bayes model, writing the log ratio of joint densities as a linear combination of those of marginal densities.

Additive models General Classification

Likelihood Adaptively Modified Penalties

no code implementations23 Aug 2013 Yang Feng, Tengfei Li, Zhiliang Ying

A new family of penalty functions, adaptive to likelihood, is introduced for model selection in general regression models.

Model Selection

APPLE: Approximate Path for Penalized Likelihood Estimators

no code implementations2 Nov 2012 Yi Yu, Yang Feng

In high-dimensional data analysis, penalized likelihood estimators are shown to provide superior results in both variable selection and parameter estimation.

Variable Selection

Cannot find the paper you are looking for? You can Submit a new open access paper.