no code implementations • SIGDIAL (ACL) 2021 • Minh Nguyen, Zhou Yu
Deployed spoken dialog systems receive user input in the form of automatic speech recognition (ASR) transcripts, and simply applying NER model trained on written text to ASR transcripts often leads to low accuracy because compared to written text, ASR transcripts lack important cues such as punctuation and capitalization.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+4
1 code implementation • Findings (EMNLP) 2021 • Jun Zhang, Yan Yang, Chencai Chen, Liang He, Zhou Yu
Recommendation dialogs require the system to build a social bond with users to gain trust and develop affinity in order to increase the chance of a successful recommendation.
1 code implementation • SIGDIAL (ACL) 2021 • Satwik Kottur, Chinnadhurai Sankar, Zhou Yu, Alborz Geramifard
Real-world conversational agents must effectively handle long conversations that span multiple contexts.
no code implementations • SIGDIAL (ACL) 2022 • Liang Qiu, Yizhou Zhao, Yuan Liang, Pan Lu, Weiyan Shi, Zhou Yu, Song-Chun Zhu
One of which is to track the agent’s mental state transition and teach the agent to make decisions guided by its value like a human.
no code implementations • ACL 2022 • Ying Xu, Dakuo Wang, Mo Yu, Daniel Ritchie, Bingsheng Yao, Tongshuang Wu, Zheng Zhang, Toby Li, Nora Bradford, Branda Sun, Tran Hoang, Yisi Sang, Yufang Hou, Xiaojuan Ma, Diyi Yang, Nanyun Peng, Zhou Yu, Mark Warschauer
Through benchmarking with QG models, we show that the QG model trained on FairytaleQA is capable of asking high-quality and more diverse questions.
no code implementations • Findings (NAACL) 2022 • Bowen Yang, Cong Han, Yu Li, Lei Zuo, Zhou Yu
In this paper, we propose a simple yet effective architecture comprising a pre-trained language model (PLM) and an item metadata encoder to integrate the recommendation and the dialog generation better.
1 code implementation • COLING 2022 • Libo Qin, Qiguang Chen, Tianbao Xie, Qian Liu, Shijue Huang, Wanxiang Che, Zhou Yu
Consistency identification in task-oriented dialog (CI-ToD) usually consists of three subtasks, aiming to identify inconsistency between current system response and current user response, dialog history and the corresponding knowledge base.
1 code implementation • NAACL 2022 • Xun Yuan, Derek Pham, Sam Davidson, Zhou Yu
Currently available grammatical error correction (GEC) datasets are compiled using essays or other long-form text written by language learners, limiting the applicability of these datasets to other domains such as informal writing and conversational dialog.
no code implementations • 18 Mar 2025 • Zhenwei Shao, Mingyang Wang, Zhou Yu, Wenwen Pan, Yan Yang, Tao Wei, Hongyuan Zhang, Ning Mao, Wei Chen, Jun Yu
Despite the success of these token pruning methods, they still suffer from two major shortcomings: (i) considerable accuracy drop due to insensitive attention signals in early layers, and (ii) limited speedup when generating long responses (e. g., 30 tokens).
1 code implementation • 26 Feb 2025 • Matthew Toles, Nikhil Balwani, Rattandeep Singh, Valentina Giulia Sartori Rodriguez, Zhou Yu
Many real-world eligibility problems, ranging from medical diagnosis to tax planning, can be mapped to decision problems expressed in natural language, wherein a model must make a binary choice based on user features.
no code implementations • 21 Feb 2025 • Hang Yuan, Christina Dan Wang, Zhou Yu
Nonlinear sufficient dimension reduction\citep{libing_generalSDR}, which constructs nonlinear low-dimensional representations to summarize essential features of high-dimensional data, is an important branch of representation learning.
no code implementations • 17 Feb 2025 • Xiao Yu, Ruize Xu, Chengyuan Xue, Jinzhong Zhang, Zhou Yu
A reliable resume-job matching system helps a company recommend suitable candidates from a pool of resumes and helps a job seeker find relevant jobs from a list of job posts.
1 code implementation • 12 Jan 2025 • Raghav Singhal, Zachary Horvitz, Ryan Teehan, Mengye Ren, Zhou Yu, Kathleen McKeown, Rajesh Ranganath
For steering text-to-image models with a human preference reward, we find that FK steering a 0. 8B parameter model outperforms a 2. 6B parameter fine-tuned model on prompt fidelity, with faster sampling and no training.
1 code implementation • 26 Dec 2024 • Shuntuo Xu, Zhou Yu
This paper investigates the connection between neural networks and sufficient dimension reduction (SDR), demonstrating that neural networks inherently perform SDR in regression tasks under appropriate rank regularizations.
no code implementations • 23 Dec 2024 • Shuyang Liu, Ruiqiu Zheng, Yunhang Shen, Ke Li, Xing Sun, Zhou Yu, Shaohui Lin
Semi-supervised learning (SSL) assumes that neighbor points lie in the same category (neighbor assumption), and points in different clusters belong to various categories (cluster assumption).
2 code implementations • 22 Oct 2024 • Li Siyan, Vethavikashini Chithrra Raghuram, Omar Khattab, Julia Hirschberg, Zhou Yu
While open-source models, hosted locally on the user's machine, alleviate some concerns, models that users can host locally are often less capable than proprietary frontier models.
no code implementations • 2 Oct 2024 • Xiao Yu, Baolin Peng, Vineeth Vajipey, Hao Cheng, Michel Galley, Jianfeng Gao, Zhou Yu
Autonomous agents have demonstrated significant potential in automating complex multistep decision-making tasks.
no code implementations • 2 Oct 2024 • Ryan Shea, Aymen Kallala, Xin Lucy Liu, Michael W. Morris, Zhou Yu
To test the effectiveness of ACE-generated feedback, we conducted a user experiment with two consecutive trials of negotiation and found that it improves negotiation performances significantly compared to a system that doesn't provide feedback and one which uses an alternative method of providing feedback.
no code implementations • 26 Sep 2024 • Ryan Shea, Zhou Yu
Despite recent advancements in AI and NLP, negotiation remains a difficult domain for AI agents.
no code implementations • 6 Sep 2024 • Xiaoyi Liu, Zhou Yu, Lianghao Tan, Yafeng Yan, Ge Shi
To further enhance classification accuracy, we developed ensemble models employing max voting, average voting, and stacking, resulting in accuracies of 0. 803, 0. 82, and 0. 83.
no code implementations • 23 Aug 2024 • Xiaoyi Liu, Zhou Yu, Lianghao Tan
Eventually, our own model, MobileNet-Lung based on MobileNetV2, with fine-tuning and an additional layer of attention within feature layers, was invented to tackle the lung disease classification task and achieved an accuracy of 0. 933.
1 code implementation • 9 Jul 2024 • Xiao Yu, Qingyang Wu, Yu Li, Zhou Yu
Alignment is a crucial step to enhance the instruction-following and conversational abilities of language models.
1 code implementation • 28 Jun 2024 • Xuanming Zhang, Anthony Diaz, Zixun Chen, Qingyang Wu, Kun Qian, Erik Voss, Zhou Yu
To bridge this gap, we introduce DECOR, a novel benchmark that includes expert annotations for detecting incoherence in L2 English writing, identifying the underlying reasons, and rewriting the incoherent sentences.
1 code implementation • 25 Jun 2024 • Li Siyan, Teresa Shao, Zhou Yu, Julia Hirschberg
Student passion and perseverance, or grit, has been associated with language learning success.
1 code implementation • 25 Jun 2024 • Kun Qian, Shunji Wan, Claudia Tang, Youzhi Wang, Xuanming Zhang, Maximillian Chen, Zhou Yu
As large language models achieve impressive scores on traditional benchmarks, an increasing number of researchers are becoming concerned about benchmark data leakage during pre-training, commonly known as the data contamination problem.
1 code implementation • 21 Jun 2024 • Zachary Horvitz, Ajay Patel, Kanishk Singh, Chris Callison-Burch, Kathleen McKeown, Zhou Yu
The goal of text style transfer is to transform the style of texts while preserving their original meaning, often with only a few examples of the target style.
no code implementations • 22 May 2024 • Xiaoyi Liu, Hongjie Qiu, Muqing Li, Zhou Yu, Yutian Yang, Yafeng Yan
This paper introduces an innovative multi-modal fusion deep learning approach to overcome the drawbacks of traditional single-modal recognition techniques.
1 code implementation • 20 May 2024 • Zhenwei Shao, Zhou Yu, Jun Yu, Xuecheng Ouyang, Lihao Zheng, Zhenbiao Gai, Mingyang Wang, Jiajun Ding
By harnessing the capabilities of large language models (LLMs), recent large multimodal models (LMMs) have shown remarkable versatility in open-world multimodal understanding.
Ranked #85 on
Visual Question Answering
on MM-Vet
no code implementations • 29 Apr 2024 • Jiajie Yuan, Linxiao Wu, Yulu Gong, Zhou Yu, Ziang Liu, Shuyao He
This paper combines Struts and Hibernate two architectures together, using DAO (Data Access Object) to store and access data.
no code implementations • Findings (ACL) 2022 • Yingwen Fu, Wenjie Ou, Zhou Yu, Yue Lin
Unsupervised constrained text generation aims to generate text under a given set of constraints without any supervised data.
2 code implementations • 21 Apr 2024 • Li Siyan, Teresa Shao, Zhou Yu, Julia Hirschberg
Existing English-teaching chatbots rarely incorporate empathy explicitly in their feedback, but empathetic feedback could help keep students engaged and reduce learner anxiety.
no code implementations • 16 Apr 2024 • Rui Qiu, Zhou Yu, Zhenhua Lin
This paper explores the field of semi-supervised Fr\'echet regression, driven by the significant costs associated with obtaining non-Euclidean labels.
no code implementations • 23 Mar 2024 • Bin Gao, Zhuomin He, Puru Sharma, Qingxuan Kang, Djordje Jevdjic, Junbo Deng, Xingkun Yang, Zhou Yu, Pengfei Zuo
Interacting with humans through multi-turn conversations is a fundamental feature of large language models (LLMs).
1 code implementation • 1 Mar 2024 • Xiao Yu, Yunan Lu, Zhou Yu
Retrieval-augmented question-answering systems combine retrieval techniques with large language models to provide answers that are more accurate and informative.
1 code implementation • 23 Feb 2024 • Zachary Horvitz, Jingru Chen, Rahul Aditya, Harshvardhan Srivastava, Robert West, Zhou Yu, Kathleen McKeown
Humor is a fundamental facet of human cognition and interaction.
1 code implementation • 19 Feb 2024 • Yanda Chen, Chen Zhao, Zhou Yu, Kathleen McKeown, He He
Pre-trained language models (LMs) are capable of in-context learning (ICL): they can adapt to a task with only a few examples given in the prompt without any parameter update.
1 code implementation • CVPR 2024 • Myeongseob Ko, Feiyang Kang, Weiyan Shi, Ming Jin, Zhou Yu, Ruoxi Jia
Inspired by this, we introduce a new method for estimating the influence of training data, which requires calculating gradients for specific test samples, paired with a forward pass for each training point.
1 code implementation • 30 Jan 2024 • Ming Gu, Yan Yang, Chengcai Chen, Zhou Yu
Experimental results on the MultiWOZ 2. 1 dataset show that our method which has only less than 1 billion parameters achieves state-of-the-art performance under the data ratio settings of 5%, 10%, and 25% when limited to models under 100 billion parameters.
no code implementations • 29 Jan 2024 • Xiao Yu, Jinzhong Zhang, Zhou Yu
A reliable resume-job matching system helps a company find suitable candidates from a pool of resumes, and helps a job seeker find relevant jobs from a list of job posts.
1 code implementation • 28 Jan 2024 • Liguo Zhou, Yinglei Song, Yichao Gao, Zhou Yu, Michael Sodamin, Hongshen Liu, Liang Ma, Lian Liu, Hao liu, Yang Liu, Haichuan Li, Guang Chen, Alois Knoll
However, the availability of free and open-source simulators is limited, and the installation and configuration process can be daunting for beginners and interdisciplinary researchers.
1 code implementation • 21 Jan 2024 • Xuanming Zhang, Zixun Chen, Zhou Yu
To bridge this gap, we propose a new task, language proficiency-oriented lexical substitution.
no code implementations • CVPR 2024 • Jimin Xu, Tianbao Wang, Tao Jin, Shengyu Zhang, Dongjie Fu, Zhe Wang, Jiangjing Lyu, Chengfei Lv, Chaoyue Niu, Zhou Yu, Zhou Zhao, Fei Wu
Specifically in the first stage MPOD123 utilizes the pretrained view-conditioned diffusion model to guide the outline shape optimization of the 3D content.
1 code implementation • 13 Dec 2023 • Yunchen Li, Zhou Yu, Gaoqi He, Yunhang Shen, Ke Li, Xing Sun, Shaohui Lin
On the other hand, the model unconditionally learns the probability distribution of the data $p(X)$ and generates samples that conform to this distribution.
no code implementations • 15 Nov 2023 • Libo Qin, Wenbo Pan, Qiguang Chen, Lizi Liao, Zhou Yu, Yue Zhang, Wanxiang Che, Min Li
End-to-end task-oriented dialogue (EToD) can directly generate responses in an end-to-end fashion without modular training, which attracts escalating popularity.
1 code implementation • 20 Oct 2023 • Xiao Yu, Baolin Peng, Michel Galley, Jianfeng Gao, Zhou Yu
The self-improving ability of large language models (LLMs), enabled by prompting them to analyze and revise their own outputs, has garnered significant interest in recent research.
no code implementations • 17 Oct 2023 • Matthew Toles, Yukun Huang, Zhou Yu, Luis Gravano
To enable evaluation of factual domain clarification question generation, We present a new task that focuses on the ability to elicit missing information in multi-hop reasoning tasks.
1 code implementation • 16 Oct 2023 • Ryan Shea, Zhou Yu
Our automatic and human evaluations show that our framework improves both the persona consistency and dialogue quality of a state-of-the-art social chatbot.
no code implementations • 28 Sep 2023 • Yu Li, Shang Qu, Jili Shen, Shangchao Min, Zhou Yu
Chatbots have become popular in educational settings, revolutionizing how students interact with material and how teachers teach.
1 code implementation • 29 Aug 2023 • Zachary Horvitz, Ajay Patel, Chris Callison-Burch, Zhou Yu, Kathleen McKeown
Our parameter-efficient approach, ParaGuide, leverages paraphrase-conditioned diffusion models alongside gradient-based guidance from both off-the-shelf classifiers and strong existing style embedders to transform the style of text while preserving semantic information.
no code implementations • 24 Aug 2023 • Yu-Wen Chen, Zhou Yu, Julia Hirschberg
Pronunciation assessment models designed for open response scenarios enable users to practice language skills in a manner similar to real-life communication.
1 code implementation • 19 Jul 2023 • JianGuo Zhang, Kun Qian, Zhiwei Liu, Shelby Heinecke, Rui Meng, Ye Liu, Zhou Yu, Huan Wang, Silvio Savarese, Caiming Xiong
Despite advancements in conversational AI, language models encounter challenges to handle diverse conversational tasks, and existing dialogue dataset collections often lack diversity and comprehensiveness.
no code implementations • 17 Jul 2023 • Yanda Chen, Ruiqi Zhong, Narutatsu Ri, Chen Zhao, He He, Jacob Steinhardt, Zhou Yu, Kathleen McKeown
To answer these questions, we propose to evaluate $\textbf{counterfactual simulatability}$ of natural language explanations: whether an explanation can enable humans to precisely infer the model's outputs on diverse counterfactuals of the explained input.
1 code implementation • 1 Jul 2023 • Shaohui Lin, Wenxuan Huang, Jiao Xie, Baochang Zhang, Yunhang Shen, Zhou Yu, Jungong Han, David Doermann
In this paper, we propose a novel Knowledge-driven Differential Filter Sampler~(KDFS) with Masked Filter Modeling~(MFM) framework for filter pruning, which globally prunes the redundant filters based on the prior knowledge of a pre-trained model in a differential and non-alternative optimization.
no code implementations • 23 May 2023 • Qingyang Wu, Deema Alnuhait, Derek Chen, Zhou Yu
We demonstrate our paradigm in practice through MultiWOZ-Remake, including an interactive textual interface built for the MultiWOZ database and a correspondingly re-processed dataset.
1 code implementation • 23 May 2023 • Xiao Yu, Maximillian Chen, Zhou Yu
Planning for goal-oriented dialogue often requires simulating future dialogue interactions and estimating task progress.
no code implementations • 23 May 2023 • Narutatsu Ri, Bill Sun, Sam Davidson, Zhou Yu
Although significant progress has been made in developing methods for Grammatical Error Correction (GEC), addressing word choice improvements has been notably lacking and enhancing sentence expressivity by replacing phrases with advanced expressions is an understudied aspect.
1 code implementation • 23 May 2023 • Sky CH-Wang, Arkadiy Saakyan, Oliver Li, Zhou Yu, Smaranda Muresan
Embedding Chain-of-Thought prompting in a human-AI collaborative framework, we build a high-quality dataset of 3, 069 social norms aligned with social situations across Chinese and American cultures alongside corresponding free-text explanations.
1 code implementation • 6 May 2023 • Maximillian Chen, Xiao Yu, Weiyan Shi, Urvi Awasthi, Zhou Yu
The standard approach has been fine-tuning pre-trained language models to perform generation conditioned on these intents.
1 code implementation • CVPR 2023 • Zhou Yu, Lixiang Zheng, Zhou Zhao, Fei Wu, Jianping Fan, Kui Ren, Jun Yu
A recent benchmark AGQA poses a promising paradigm to generate QA pairs automatically from pre-annotated scene graphs, enabling it to measure diverse reasoning abilities with granular control.
no code implementations • 11 Apr 2023 • Kun Qian, Ryan Shea, Yu Li, Luke Kutszik Fryer, Zhou Yu
Along with the development of systems for natural language understanding and generation, dialog systems have been widely adopted for language learning and practicing.
1 code implementation • 15 Mar 2023 • Rahul Goel, Waleed Ammar, Aditya Gupta, Siddharth Vashishtha, Motoki Sano, Faiz Surani, Max Chang, HyunJeong Choe, David Greene, Kyle He, Rattima Nitisaroj, Anna Trukhina, Shachi Paul, Pararth Shah, Rushin Shah, Zhou Yu
Research interest in task-oriented dialogs has increased as systems such as Google Assistant, Alexa and Siri have become ubiquitous in everyday life.
1 code implementation • 8 Mar 2023 • Deema Alnuhait, Qingyang Wu, Zhou Yu
While current dialogue systems like ChatGPT have made significant advancements in text-based interactions, they often overlook the potential of other modalities in enhancing the overall user experience.
1 code implementation • CVPR 2023 • Zhou Yu, Xuecheng Ouyang, Zhenwei Shao, Meng Wang, Jun Yu
Knowledge-based visual question answering (VQA) requires external knowledge beyond the image to answer the question.
Ranked #3 on
Visual Question Answering (VQA)
on A-OKVQA
1 code implementation • 2 Mar 2023 • Derek Chen, Celine Lee, Yunan Lu, Domenic Rosati, Zhou Yu
Large language models (LLMs) effectively generate fluent text when the target output follows natural language patterns.
1 code implementation • 24 Feb 2023 • Maximillian Chen, Zhou Yu
Speech models have long been known to overfit individual speakers for many classification tasks.
no code implementations • 24 Feb 2023 • Baolin Peng, Michel Galley, Pengcheng He, Hao Cheng, Yujia Xie, Yu Hu, Qiuyuan Huang, Lars Liden, Zhou Yu, Weizhu Chen, Jianfeng Gao
Large language models (LLMs), such as ChatGPT, are able to generate human-like, fluent responses for many downstream tasks, e. g., task-oriented dialog and question answering.
no code implementations • 12 Feb 2023 • Derek Chen, Kun Qian, Zhou Yu
Prompt-based methods with large pre-trained language models (PLMs) have shown impressive unaided performance across many NLP tasks.
1 code implementation • 7 Feb 2023 • Maximillian Chen, Alexandros Papangelis, Chenyang Tao, Seokhwan Kim, Andy Rosenbaum, Yang Liu, Zhou Yu, Dilek Hakkani-Tur
Collecting high quality conversational data can be very expensive for most applications and infeasible for others due to privacy, ethical, or similar concerns.
no code implementations • 20 Dec 2022 • Yu Li, Baolin Peng, Pengcheng He, Michel Galley, Zhou Yu, Jianfeng Gao
In this work, we propose DIONYSUS (dynamic input optimization in pre-training for dialogue summarization), a pre-trained encoder-decoder model for summarizing dialogues in any new domain.
no code implementations • 20 Dec 2022 • Yukun Huang, Yanda Chen, Zhou Yu, Kathleen McKeown
We propose to combine in-context learning objectives with language modeling objectives to distill both the ability to read in-context examples and task knowledge to the smaller models.
no code implementations • 19 Dec 2022 • Yingwen Fu, Wenjie Ou, Zhou Yu, Yue Lin
Conversational text-to-SQL is designed to translate multi-turn natural language questions into their corresponding SQL queries.
no code implementations • 6 Dec 2022 • Derek Chen, Zhou Yu
Training dialogue systems often entails dealing with noisy training examples and unexpected user inputs.
no code implementations • 1 Dec 2022 • Mingyang Zhou, Grace Luo, Anna Rohrbach, Zhou Yu
In our paper, we first demonstrate that by combining more fine-grained context that captures the key named entities (obtained via an oracle) and the global context that summarizes the news, we can dramatically improve the model's ability to generate accurate news captions.
1 code implementation • 30 Nov 2022 • Xiao Yu, Qingyang Wu, Kun Qian, Zhou Yu
In task-oriented dialogs (TOD), reinforcement learning (RL) algorithms train a model to directly optimize response for task-related metrics.
no code implementations • 22 Nov 2022 • Weiyan Shi, Emily Dinan, Adi Renduchintala, Daniel Fried, Athul Paul Jacob, Zhou Yu, Mike Lewis
Existing approaches built separate classifiers to detect nonsense in dialogues.
no code implementations • 25 Oct 2022 • Maximillian Chen, Alexandros Papangelis, Chenyang Tao, Andy Rosenbaum, Seokhwan Kim, Yang Liu, Zhou Yu, Dilek Hakkani-Tur
Dialogue understanding tasks often necessitate abundant annotated data to achieve good performance and that presents challenges in low-resource settings.
1 code implementation • 22 Oct 2022 • David Gros, Yu Li, Zhou Yu
Dialog systems are often designed or trained to output human-like responses.
1 code implementation • 17 Oct 2022 • Sky CH-Wang, Evan Li, Oliver Li, Smaranda Muresan, Zhou Yu
Affective responses to music are highly personal.
no code implementations • 11 Oct 2022 • Kushal Chawla, Weiyan Shi, Jingwen Zhang, Gale Lucas, Zhou Yu, Jonathan Gratch
Dialogue systems capable of social influence such as persuasion, negotiation, and therapy, are essential for extending the use of technology to numerous realistic scenarios.
1 code implementation • 16 Sep 2022 • Yanda Chen, Chen Zhao, Zhou Yu, Kathleen McKeown, He He
In-context learning (ICL) suffers from oversensitivity to the prompt, making it unreliable in real-world scenarios.
1 code implementation • 15 Sep 2022 • Qingyang Wu, Zhou Yu
Transformer encoder-decoder models have achieved great performance in dialogue generation tasks, however, their inability to process long dialogue history often leads to truncation of the context To address this problem, we propose a novel memory-augmented transformer that is compatible with existing pre-trained encoder-decoder models and enables efficient preservation of the dialogue history information.
no code implementations • 31 Jul 2022 • Yu Li, Chun-Yen Chen, Dian Yu, Sam Davidson, Ryan Hou, Xun Yuan, Yinghua Tan, Derek Pham, Zhou Yu
This paper reports on progress towards building an online language learning tool to provide learners with conversational experience by using dialog systems as conversation practice partners.
1 code implementation • 22 Jun 2022 • Baolin Peng, Michel Galley, Pengcheng He, Chris Brockett, Lars Liden, Elnaz Nouri, Zhou Yu, Bill Dolan, Jianfeng Gao
We introduce GODEL (Grounded Open Dialogue Language Model), a large pre-trained language model for dialog.
no code implementations • 25 May 2022 • Yukun Huang, Kun Qian, Zhou Yu
So pre-trained prompt tuning (PPT) is proposed to initialize prompts by leveraging pre-training data.
1 code implementation • 15 Apr 2022 • Weiyan Shi, Ryan Shea, Si Chen, Chiyuan Zhang, Ruoxi Jia, Zhou Yu
Utilizing the fact that sensitive information in language data tends to be sparse, Shi et al. (2021) formalized a DP notion extension called Selective Differential Privacy (SDP) to protect only the sensitive tokens defined by a policy function.
no code implementations • 26 Mar 2022 • Sha Yuan, Hanyu Zhao, Shuai Zhao, Jiahong Leng, Yangxiao Liang, Xiaozhi Wang, Jifan Yu, Xin Lv, Zhou Shao, Jiaao He, Yankai Lin, Xu Han, Zhenghao Liu, Ning Ding, Yongming Rao, Yizhao Gao, Liang Zhang, Ming Ding, Cong Fang, Yisen Wang, Mingsheng Long, Jing Zhang, Yinpeng Dong, Tianyu Pang, Peng Cui, Lingxiao Huang, Zheng Liang, HuaWei Shen, HUI ZHANG, Quanshi Zhang, Qingxiu Dong, Zhixing Tan, Mingxuan Wang, Shuo Wang, Long Zhou, Haoran Li, Junwei Bao, Yingwei Pan, Weinan Zhang, Zhou Yu, Rui Yan, Chence Shi, Minghao Xu, Zuobai Zhang, Guoqiang Wang, Xiang Pan, Mengjie Li, Xiaoyu Chu, Zijun Yao, Fangwei Zhu, Shulin Cao, Weicheng Xue, Zixuan Ma, Zhengyan Zhang, Shengding Hu, Yujia Qin, Chaojun Xiao, Zheni Zeng, Ganqu Cui, Weize Chen, Weilin Zhao, Yuan YAO, Peng Li, Wenzhao Zheng, Wenliang Zhao, Ziyi Wang, Borui Zhang, Nanyi Fei, Anwen Hu, Zenan Ling, Haoyang Li, Boxi Cao, Xianpei Han, Weidong Zhan, Baobao Chang, Hao Sun, Jiawen Deng, Chujie Zheng, Juanzi Li, Lei Hou, Xigang Cao, Jidong Zhai, Zhiyuan Liu, Maosong Sun, Jiwen Lu, Zhiwu Lu, Qin Jin, Ruihua Song, Ji-Rong Wen, Zhouchen Lin, LiWei Wang, Hang Su, Jun Zhu, Zhifang Sui, Jiajun Zhang, Yang Liu, Xiaodong He, Minlie Huang, Jian Tang, Jie Tang
With the rapid development of deep learning, training Big Models (BMs) for multiple downstream tasks becomes a popular paradigm.
1 code implementation • 26 Mar 2022 • Ying Xu, Dakuo Wang, Mo Yu, Daniel Ritchie, Bingsheng Yao, Tongshuang Wu, Zheng Zhang, Toby Jia-Jun Li, Nora Bradford, Branda Sun, Tran Bao Hoang, Yisi Sang, Yufang Hou, Xiaojuan Ma, Diyi Yang, Nanyun Peng, Zhou Yu, Mark Warschauer
Through benchmarking with QG models, we show that the QG model trained on FairytaleQA is capable of asking high-quality and more diverse questions.
Ranked #1 on
Question Generation
on FairytaleQA
1 code implementation • 24 Mar 2022 • Zhou Yu, Zitian Jin, Jun Yu, Mingliang Xu, Hongbo Wang, Jianping Fan
Recent advances in Transformer architectures [1] have brought remarkable improvements to visual question answering (VQA).
no code implementations • 18 Mar 2022 • Shikib Mehri, Jinho Choi, Luis Fernando D'Haro, Jan Deriu, Maxine Eskenazi, Milica Gasic, Kallirroi Georgila, Dilek Hakkani-Tur, Zekang Li, Verena Rieser, Samira Shaikh, David Traum, Yi-Ting Yeh, Zhou Yu, Yizhe Zhang, Chen Zhang
This is a report on the NSF Future Directions Workshop on Automatic Evaluation of Dialog.
1 code implementation • 15 Mar 2022 • Maximillian Chen, Caitlyn Chen, Xiao Yu, Zhou Yu
Syntax is a fundamental component of language, yet few metrics have been employed to capture syntactic similarity or coherence at the utterance- and document-level.
no code implementations • 15 Mar 2022 • Maximillian Chen, Weiyan Shi, Feifan Yan, Ryan Hou, Jingwen Zhang, Saurav Sahay, Zhou Yu
Complex conversation settings such as persuasion involve communicating changes in attitude or behavior, so users' perspectives need to be addressed, even when not directly related to the topic.
no code implementations • CVPR 2022 • Mingyang Zhou, Licheng Yu, Amanpreet Singh, Mengjiao Wang, Zhou Yu, Ning Zhang
We adapt our pre-trained model to a set of V+L downstream tasks, including VQA, NLVR2, Visual Entailment, and RefCOCO+.
no code implementations • 10 Feb 2022 • Rui Qiu, Zhou Yu, Ruoqing Zhu
Statistical analysis is increasingly confronted with complex data from metric spaces.
no code implementations • 24 Dec 2021 • Chaoxia Yuan, Chao Ying, Zhou Yu, Fang Fang
Support vector machine (SVM) is a powerful classification method that has achieved great success in many fields.
no code implementations • 15 Dec 2021 • Mingyang Zhou, Mahasweta Chakraborti, Sijia Qian, Zhou Yu, Jingwen Zhang
The dataset and classifiers contribute to monitoring and tracking of vaccine discussions for social scientific and public health efforts in combating the problem of vaccine misinformation.
1 code implementation • 15 Dec 2021 • Bowen Yang, Cong Han, Yu Li, Lei Zuo, Zhou Yu
The encoder learns to map item metadata to embeddings that can reflect the semantic information in the dialog context.
no code implementations • 15 Dec 2021 • Lei Zuo, Kun Qian, Bowen Yang, Zhou Yu
A commonly observed problem of the state-of-the-art natural language technologies, such as Amazon Alexa and Apple Siri, is that their services do not extend to most developing countries' citizens due to language barriers.
no code implementations • NAACL 2022 • Kun Qian, Ahmad Beirami, Satwik Kottur, Shahin Shayandeh, Paul Crook, Alborz Geramifard, Zhou Yu, Chinnadhurai Sankar
We find that training on our augmented dialog data improves the model's ability to deal with ambiguous scenarios, without sacrificing performance on unmodified turns.
1 code implementation • 15 Dec 2021 • Xun Yuan, Derek Pham, Sam Davidson, Zhou Yu
Currently available grammatical error correction (GEC) datasets are compiled using well-formed written text, limiting the applicability of these datasets to other domains such as informal writing and dialog.
no code implementations • SIGDIAL (ACL) 2022 • Qingyang Wu, Song Feng, Derek Chen, Sachindra Joshi, Luis A. Lastras, Zhou Yu
Collecting data for training dialog systems can be extremely expensive due to the involvement of human participants and need for extensive annotation.
no code implementations • NAACL 2022 • Yu Li, Baolin Peng, Yelong Shen, Yi Mao, Lars Liden, Zhou Yu, Jianfeng Gao
To address these challenges, we present PLUG, a language model that homogenizes different knowledge sources to a unified knowledge representation for knowledge-grounded dialogue generation tasks.
1 code implementation • 25 Oct 2021 • Pan Lu, Liang Qiu, Jiaqi Chen, Tony Xia, Yizhou Zhao, Wei zhang, Zhou Yu, Xiaodan Liang, Song-Chun Zhu
Also, we develop a strong IconQA baseline Patch-TRM that applies a pyramid cross-modal Transformer with input diagram embeddings pre-trained on the icon dataset.
Ranked #1 on
Visual Question Answering (VQA)
on IconQA
no code implementations • DeepLo 2022 • Derek Chen, Zhou Yu, Samuel R. Bowman
Crowdsourcing platforms are often used to collect datasets for training machine learning models, despite higher levels of inaccurate labeling compared to expert labeling.
1 code implementation • EMNLP 2021 • Zhaojiang Lin, Bing Liu, Andrea Madotto, Seungwhan Moon, Paul Crook, Zhenpeng Zhou, Zhiguang Wang, Zhou Yu, Eunjoon Cho, Rajen Subba, Pascale Fung
Zero-shot transfer learning for dialogue state tracking (DST) enables us to handle a variety of task-oriented dialogue domains without the expense of collecting in-domain data.
2 code implementations • EMNLP 2021 • Derek Chen, Zhou Yu
Practical dialogue systems require robust methods of detecting out-of-scope (OOS) utterances to avoid conversational breakdowns and related failure modes.
1 code implementation • NAACL 2022 • Weiyan Shi, Aiqi Cui, Evan Li, Ruoxi Jia, Zhou Yu
Given that the private information in natural language is sparse (for example, the bulk of an email might not carry personally identifiable information), we propose a new privacy notion, selective differential privacy, to provide rigorous privacy guarantees on the sensitive portion of the data to improve model utility.
1 code implementation • 16 Aug 2021 • Yuhao Cui, Zhou Yu, Chunqi Wang, Zhongzhou Zhao, Ji Zhang, Meng Wang, Jun Yu
Nevertheless, most existing VLP approaches have not fully utilized the intrinsic knowledge within the image-text pairs, which limits the effectiveness of the learned alignments and further restricts the performance of their models.
no code implementations • ACL 2021 • Meng Zhou, Zechen Li, Bowen Tan, Guangtao Zeng, Wenmian Yang, Xuehai He, Zeqian Ju, Subrato Chakravorty, Shu Chen, Xingyi Yang, Yichen Zhang, Qingyang Wu, Zhou Yu, Kun Xu, Eric Xing, Pengtao Xie
Training complex dialog generation models on small datasets bears high risk of overfitting.
no code implementations • ACL 2021 • Vojt{\v{e}}ch Hude{\v{c}}ek, Ond{\v{r}}ej Du{\v{s}}ek, Zhou Yu
Our model demonstrates state-of-the-art performance in slot tagging without labeled training data on four different dialogue domains.
no code implementations • ACL 2021 • Jing Gu, Qingyang Wu, Chongruo wu, Weiyan Shi, Zhou Yu
However, the performance of pre-trained models on task-oriented dialog tasks is still under-explored.
1 code implementation • SIGDIAL (ACL) 2021 • Kai-Hui Liang, Patrick Lange, Yoo Jung Oh, Jingwen Zhang, Yoshimi Fukuoka, Zhou Yu
To develop intervention chatbots, the first step is to understand natural language conversation strategies in human conversation.
no code implementations • ACL 2021 • David Gros, Yu Li, Zhou Yu
Humans are increasingly interacting with machines through language, sometimes in contexts where the user may not know they are talking to a machine (like over the phone or a text chatbot).
no code implementations • 3 Jun 2021 • Kai-Hui Liang, Weiyan Shi, Yoojung Oh, Hao-Chuan Wang, Jingwen Zhang, Zhou Yu
Using chatbots to deliver recommendations is increasingly popular.
1 code implementation • ACL 2021 • Siyang Liu, Chujie Zheng, Orianna Demasi, Sahand Sabour, Yu Li, Zhou Yu, Yong Jiang, Minlie Huang
Emotional support is a crucial ability for many conversation scenarios, including social interactions, mental health support, and customer service chats.
no code implementations • ACL 2021 • Liang Qiu, Yuan Liang, Yizhou Zhao, Pan Lu, Baolin Peng, Zhou Yu, Ying Nian Wu, Song-Chun Zhu
Inferring social relations from dialogues is vital for building emotionally intelligent robots to interpret human language better and act accordingly.
Ranked #5 on
Dialog Relation Extraction
on DialogRE
1 code implementation • ACL 2021 • Weixin Liang, Kai-Hui Liang, Zhou Yu
Open-domain dialog systems have a user-centric goal: to provide humans with an engaging conversation experience.
1 code implementation • NAACL 2021 • Zhaojiang Lin, Bing Liu, Seungwhan Moon, Paul Crook, Zhenpeng Zhou, Zhiguang Wang, Zhou Yu, Andrea Madotto, Eunjoon Cho, Rajen Subba
Zero-shot cross-domain dialogue state tracking (DST) enables us to handle unseen domains without the expense of collecting in-domain data.
no code implementations • SIGDIAL (ACL) 2021 • Kun Qian, Ahmad Beirami, Zhouhan Lin, Ankita De, Alborz Geramifard, Zhou Yu, Chinnadhurai Sankar
In this work, we identify an overlooked issue with dialog state annotation inconsistencies in the dataset, where a slot type is tagged inconsistently across similar dialogs leading to confusion for DST modeling.
2 code implementations • 10 May 2021 • Zhaojiang Lin, Bing Liu, Seungwhan Moon, Paul Crook, Zhenpeng Zhou, Zhiguang Wang, Zhou Yu, Andrea Madotto, Eunjoon Cho, Rajen Subba
Zero-shot cross-domain dialogue state tracking (DST) enables us to handle task-oriented dialogue in unseen domains without the expense of collecting in-domain data.
1 code implementation • ACL 2021 • Yu Li, Josh Arnold, Feifan Yan, Weiyan Shi, Zhou Yu
We present LEGOEval, an open-source toolkit that enables researchers to easily evaluate dialogue systems in a few lines of code using the online crowdsource platform, Amazon Mechanical Turk.
1 code implementation • 18 Apr 2021 • Emily Sheng, Josh Arnold, Zhou Yu, Kai-Wei Chang, Nanyun Peng
Dialogue systems in the form of chatbots and personal assistants are being increasingly integrated into people's lives.
no code implementations • 16 Apr 2021 • Yu Li, Shirley Anugrah Hayati, Weiyan Shi, Zhou Yu
It is important for sociable recommendation dialog systems to perform as both on-task content and social content to engage users and gain their favor.
no code implementations • 6 Apr 2021 • Kun Qian, Wei Wei, Zhou Yu
The most recent researches on domain adaption focus on giving the model a better initialization, rather than optimizing the adaptation process.
no code implementations • EACL 2021 • Dian Yu, Zhou Yu
To validate our scheme, we leveraged transfer learning methods to train a multi-label dialog act prediction model and reached an F1 score of 0. 79.
2 code implementations • NAACL 2021 • Derek Chen, Howard Chen, Yi Yang, Alex Lin, Zhou Yu
Existing goal-oriented dialogue datasets focus mainly on identifying slots and values.
no code implementations • CVPR 2021 • Mingyang Zhou, Luowei Zhou, Shuohang Wang, Yu Cheng, Linjie Li, Zhou Yu, Jingjing Liu
Vision-and-language pre-training has achieved impressive success in learning multimodal representations between vision and language.
1 code implementation • Findings (EMNLP) 2021 • Dian Yu, Zhou Yu, Kenji Sagae
Large language models benefit from training with a large amount of unlabeled text, which gives them increasingly fluent and diverse generation capabilities.
no code implementations • 12 Mar 2021 • Liang Qiu, Yizhou Zhao, Yuan Liang, Pan Lu, Weiyan Shi, Zhou Yu, Song-Chun Zhu
One of which is to track the agent's mental state transition and teach the agent to make decisions guided by its value like a human.
1 code implementation • EACL 2021 • Jing Gu, Mostafa Mirshekari, Zhou Yu, Aaron Sisto
Conversational systems enable numerous valuable applications, and question-answering is an important component underlying many of these.
no code implementations • 1 Jan 2021 • Weiyan Shi, Yu Li, Saurav Sahay, Zhou Yu
Despite the recent success of large-scale language models on various downstream NLP tasks, the repetition and inconsistency problems still persist in dialogue response generation.
no code implementations • 1 Jan 2021 • Jian Huang, Yuling Jiao, Xu Liao, Jin Liu, Zhou Yu
We provide strong statistical guarantees for the learned representation by establishing an upper bound on the excess error of the objective function and show that it reaches the nonparametric minimax rate under mild conditions.
no code implementations • Findings (EMNLP) 2021 • Weiyan Shi, Yu Li, Saurav Sahay, Zhou Yu
Persuasion dialogue systems reflect the machine's ability to make strategic moves beyond verbal communication, and therefore differentiate themselves from task-oriented or open-domain dialogue systems and have their own unique values.
1 code implementation • EMNLP 2021 • Andrea Madotto, Zhaojiang Lin, Zhenpeng Zhou, Seungwhan Moon, Paul Crook, Bing Liu, Zhou Yu, Eunjoon Cho, Zhiguang Wang
Continual learning in task-oriented dialogue systems can allow us to add new domains and functionalities through time without incurring the high cost of a whole system retraining.
no code implementations • 17 Nov 2020 • Kaihui Liang, Austin Chau, Yu Li, Xueyuan Lu, Dian Yu, Mingyang Zhou, Ishan Jain, Sam Davidson, Josh Arnold, Minh Nguyen, Zhou Yu
Gunrock 2. 0 is built on top of Gunrock with an emphasis on user adaptation.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Orianna Demasi, Yu Li, Zhou Yu
Suicide prevention hotline counselors aid individuals during difficult times through millions of calls and chats.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Youzhi Tian, Weiyan Shi, Chen Li, Zhou Yu
Furthermore, we analyze the relationships between persuasion strategies and persuasion resistance strategies.
no code implementations • 14 Oct 2020 • Qingyang Wu, Zhenzhong Lan, Kun Qian, Jing Gu, Alborz Geramifard, Zhou Yu
Transformers have reached remarkable success in sequence modeling.
no code implementations • 3 Oct 2020 • David Gros, Hariharan Sezhiyan, Prem Devanbu, Zhou Yu
We carefully examine the underlying assumption here: that the task of generating comments sufficiently resembles the task of translating between natural languages, and so similar models and evaluation metrics could be used.
1 code implementation • EMNLP 2020 • Shirley Anugrah Hayati, Dongyeop Kang, Qingxiaoyang Zhu, Weiyan Shi, Zhou Yu
To better understand how humans make recommendations in communication, we design an annotation scheme related to recommendation strategies based on social science theories and annotate these dialogs.
no code implementations • EMNLP 2020 • Weixin Liang, James Zou, Zhou Yu
We propose Active Learning with Contrastive Explanations (ALICE), an expert-in-the-loop training framework that utilizes contrastive natural language explanations to improve data efficiency in learning.
1 code implementation • EMNLP 2020 • Liang Qiu, Yizhou Zhao, Weiyan Shi, Yuan Liang, Feng Shi, Tao Yuan, Zhou Yu, Song-Chun Zhu
Inducing a meaningful structural representation from one or a set of dialogues is a crucial but challenging task in computational linguistics.
no code implementations • 7 Aug 2020 • Jing Gu, Qingyang Wu, Zhou Yu
Automatic evaluation for open-ended natural language generation tasks remains a challenge.
no code implementations • ACL 2020 • Jiaying Hu, Yan Yang, Chencai Chen, Liang He, Zhou Yu
Dialogue state tracker is responsible for inferring user intentions through dialogue history.
1 code implementation • 10 Jun 2020 • Jian Huang, Yuling Jiao, Xu Liao, Jin Liu, Zhou Yu
We propose a deep dimension reduction approach to learning representations with these characteristics.
1 code implementation • ACL 2020 • Weixin Liang, James Zou, Zhou Yu
Our experiments show that CMADE achieves 89. 2% accuracy in the dialog comparison task.
no code implementations • 11 May 2020 • Wenmian Yang, Guangtao Zeng, Bowen Tan, Zeqian Ju, Subrato Chakravorty, Xuehai He, Shu Chen, Xingyi Yang, Qingyang Wu, Zhou Yu, Eric Xing, Pengtao Xie
On these two datasets, we train several dialogue generation models based on Transformer, GPT, and BERT-GPT.
1 code implementation • 25 Apr 2020 • Zhou Yu, Yuhao Cui, Jun Yu, Meng Wang, DaCheng Tao, Qi Tian
Most existing works focus on a single task and design neural architectures manually, which are highly task-specific and hard to generalize to different tasks.
Ranked #18 on
Visual Question Answering (VQA)
on VQA v2 test-std
no code implementations • Findings of the Association for Computational Linguistics 2020 • Jing Gu, Zhou Yu
We pro-pose a data annealing transfer learning procedure to bridge the performance gap on informal natural language understanding tasks.
1 code implementation • 24 Apr 2020 • Jing Gu, Qingyang Wu, Chongruo wu, Weiyan Shi, Zhou Yu
The recent success of large pre-trained language models such as BERT and GPT-2 has suggested the effectiveness of incorporating language priors in downstream dialog generation tasks.
1 code implementation • ACL 2020 • Silin Gao, Yichi Zhang, Zhijian Ou, Zhou Yu
Neural generative models have achieved promising performance on dialog generation tasks if given a huge data set.
no code implementations • 7 Apr 2020 • Qingyang Wu, Lei LI, Zhou Yu
Generative Adversarial Networks (GANs) for text generation have recently received many criticisms, as they perform worse than their MLE counterparts.
no code implementations • 16 Mar 2020 • Yijun Song, Jingwen Wang, Lin Ma, Zhou Yu, Jun Yu
The task of temporally grounding textual queries in videos is to localize one video segment that semantically corresponds to the given query.
no code implementations • 9 Mar 2020 • Xixi Zhou, Chengxi Li, Jiajun Bu, Chengwei Yao, Keyue Shi, Zhi Yu, Zhou Yu
Our approach, Text matching with Deep Info Max (TIM), is integrated with a procedure of unsupervised learning of representations by maximizing the mutual information between text matching neural network's input and output.
no code implementations • 28 Nov 2019 • Jun Jin, Chao Ying, Zhou Yu
The principal support vector machines method (Li et al., 2011) is a powerful tool for sufficient dimension reduction that replaces original predictors with their low-dimensional linear combinations without loss of information.
1 code implementation • 25 Nov 2019 • Yu Li, Kun Qian, Weiyan Shi, Zhou Yu
End-to-end task-oriented dialog models have achieved promising performance on collaborative tasks where users willingly coordinate with the system to complete a given task.
no code implementations • 25 Nov 2019 • Xiyuan Zhang, Chengxi Li, Dian Yu, Samuel Davidson, Zhou Yu
We then train a prediction model using both utterances containing ellipsis and our automatically completed utterances.
no code implementations • 25 Nov 2019 • Qingyang Wu, Lei LI, Hao Zhou, Ying Zeng, Zhou Yu
We propose to automate this headline editing process through neural network models to provide more immediate writing support for these social media news writers.
6 code implementations • 24 Nov 2019 • Yichi Zhang, Zhijian Ou, Zhou Yu
Conversations have an intrinsic one-to-many property, which means that multiple responses can be appropriate for the same dialog context.
Ranked #6 on
End-To-End Dialogue Modelling
on MULTIWOZ 2.0
1 code implementation • EACL 2021 • Qingyang Wu, Yichi Zhang, Yu Li, Zhou Yu
Existing dialog system models require extensive human annotations and are difficult to generalize to different tasks.
no code implementations • IJCNLP 2019 • Dian Yu, Michelle Cohn, Yi Mang Yang, Chun-Yen Chen, Weiming Wen, Jiaping Zhang, Mingyang Zhou, Kevin Jesse, Austin Chau, Antara Bhowmick, Shreenath Iyer, Giritheja Sreenivasulu, Sam Davidson, Ashwin Bhandare, Zhou Yu
Gunrock is the winner of the 2018 Amazon Alexa Prize, as evaluated by coherence and engagement from both real users and Amazon-selected expert conversationalists.
1 code implementation • ICLR 2020 • Lin Yuan, Zhou Yu
The demand for abstractive dialog summary is growing in real-world applications.
no code implementations • ICLR 2020 • Yiheng Zhou, Yulia Tsvetkov, Alan W. black, Zhou Yu
We train FSTs on a set of strategies and tactics used in negotiation dialogs.
2 code implementations • 12 Sep 2019 • Weixin Liang, Youzhi Tian, Chengcai Chen, Zhou Yu
To utilize limited training data more efficiently, we propose Modular Supervision Network (MOSS), an encoder-decoder training framework that could incorporate supervision from various intermediate dialog system modules including natural language understanding, dialog state tracking, dialog policy learning, and natural language generation.
no code implementations • IJCNLP 2019 • Sam Davidson, Dian Yu, Zhou Yu
Dependency parsing of conversational input can play an important role in language understanding for dialog systems by identifying the relationships between entities extracted from user utterances.
no code implementations • IJCNLP 2019 • Mingyang Zhou, Josh Arnold, Zhou Yu
Reinforcement learning (RL) is an effective approach to learn an optimal dialog policy for task-oriented visual dialog systems.
1 code implementation • IJCNLP 2019 • Weiyan Shi, Kun Qian, Xuewei Wang, Zhou Yu
We propose a method of standardizing user simulator building that can be used by the community to compare dialog system quality using the same set of user simulators fairly.
no code implementations • WS 2019 • Michelle Cohn, Chun-Yen Chen, Zhou Yu
This study tests the effect of cognitive-emotional expression in an Alexa text-to-speech (TTS) voice on users{'} experience with a social dialog system.
1 code implementation • 27 Aug 2019 • Dian Yu, Zhou Yu
Previous dialog act schemes, such as SWBD-DAMSL, are designed for human-human conversations, in which conversation partners have perfect language understanding ability.
no code implementations • 12 Aug 2019 • Zhou Yu, Yuhao Cui, Jun Yu, DaCheng Tao, Qi Tian
Learning an effective attention mechanism for multimodal data is important in many vision-and-language tasks that require a synergic understanding of both the visual and textual contents.
7 code implementations • CVPR 2019 • Zhou Yu, Jun Yu, Yuhao Cui, DaCheng Tao, Qi Tian
In this paper, we propose a deep Modular Co-Attention Network (MCAN) that consists of Modular Co-Attention (MCA) layers cascaded in depth.
Ranked #7 on
Question Answering
on SQA3D
3 code implementations • ACL 2019 • Xuewei Wang, Weiyan Shi, Richard Kim, Yoojung Oh, Sijia Yang, Jingwen Zhang, Zhou Yu
Developing intelligent persuasive conversational agents to change people's opinions and actions for social good is the frontier in advancing the ethical development of automated dialogue systems.
no code implementations • 11 Jun 2019 • Qingyang Wu, He Li, Lexin Li, Zhou Yu
With the widespread success of deep neural networks in science and technology, it is becoming increasingly important to quantify the uncertainty of the predictions produced by deep learning.
1 code implementation • ACL 2019 • Kun Qian, Zhou Yu
We train a dialog system model using multiple rich-resource single-domain dialog data by applying the model-agnostic meta-learning algorithm to dialog domain.
1 code implementation • 6 Jun 2019 • Zhou Yu, Dejing Xu, Jun Yu, Ting Yu, Zhou Zhao, Yueting Zhuang, DaCheng Tao
It is both crucial and natural to extend this research direction to the video domain for video question answering (VideoQA).
Ranked #31 on
Video Question Answering
on ActivityNet-QA
Visual Question Answering (VQA)
Zero-Shot Video Question Answer
no code implementations • 20 May 2019 • Jun Yu, Jing Li, Zhou Yu, Qingming Huang
Despite the success of existing studies, current methods only model the co-attention that characterizes the inter-modal interactions while neglecting the self-attention that characterizes the intra-modal interactions.
no code implementations • 16 Apr 2019 • Jun Yu, Jinghan Yao, Jian Zhang, Zhou Yu, DaCheng Tao
In this paper, we propose a one-stage framework, SPRNet, which performs efficient instance segmentation by introducing a single pixel reconstruction (SPR) branch to off-the-shelf one-stage detectors.
1 code implementation • NAACL 2019 • Weiyan Shi, Tiancheng Zhao, Zhou Yu
The learned dialog structure can shed light on how to analyze human dialogs, and more importantly contribute to the design and evaluation of dialog systems.
no code implementations • 1 Nov 2018 • Jiaao Chen, Jianshu Chen, Zhou Yu
The ability to select an appropriate story ending is the first step towards perfect narrative comprehension.
2 code implementations • 15 Oct 2018 • Youzhi Tian, Zhiting Hu, Zhou Yu
Text style transfer aims to modify the style of a sentence while keeping its content unchanged.
1 code implementation • EMNLP 2018 • Mingyang Zhou, Runxiang Cheng, Yong Jae Lee, Zhou Yu
The model leverages a visual attention grounding mechanism that links the visual semantics with the corresponding textual semantics.
Ranked #12 on
Multimodal Machine Translation
on Multi30K
2 code implementations • EMNLP 2018 • Weiming Wen, Songwen Su, Zhou Yu
With the increasing popularity of smart devices, rumors with multimedia content become more and more common on social networks.
no code implementations • 23 Jun 2018 • Ning Zhang, Zhou Yu, Qiang Wu
The new algorithm, called overlapping sliced inverse regression (OSIR), is able to estimate the effective dimension reduction space and determine the number of effective factors more accurately.
1 code implementation • 9 May 2018 • Zhou Yu, Jun Yu, Chenchao Xiang, Zhou Zhao, Qi Tian, DaCheng Tao
Visual grounding aims to localize an object in an image referred to by a textual query phrase.
Ranked #9 on
Phrase Grounding
on Flickr30k Entities Test
no code implementations • WS 2018 • Jiaping Zhang, Tiancheng Zhao, Zhou Yu
We propose a multimodal hierarchical reinforcement learning framework that dynamically integrates vision and language for task-oriented visual dialog.
Hierarchical Reinforcement Learning
reinforcement-learning
+4
no code implementations • ACL 2018 • Weiyan Shi, Zhou Yu
End-to-end learning framework is useful for building dialog systems for its simplicity in training and efficiency in model updating.
2 code implementations • 10 Aug 2017 • Zhou Yu, Jun Yu, Chenchao Xiang, Jianping Fan, DaCheng Tao
For fine-grained image and question representations, a `co-attention' mechanism is developed by using a deep neural network architecture to jointly learn the attentions for both the image and the question, which can allow us to reduce the irrelevant features effectively and obtain more discriminative features for image and question representations.
6 code implementations • ICCV 2017 • Zhou Yu, Jun Yu, Jianping Fan, DaCheng Tao
For multi-modal feature fusion, here we develop a Multi-modal Factorized Bilinear (MFB) pooling approach to efficiently and effectively combine multi-modal features, which results in superior performance for VQA compared with other bilinear pooling approaches.
no code implementations • 1 Mar 2017 • Zhou Yu, Alan W. black, Alexander I. Rudnicky
These systems work well when users have clear and explicit intentions that are well-aligned to the systems' capabilities.