1 code implementation • 16 Oct 2023 • Weixiao Zhou, Gengyao Li, Xianfu Cheng, Xinnian Liang, Junnan Zhu, FeiFei Zhai, Zhoujun Li
Specifically, we first conduct domain-aware pre-training using large-scale multi-scenario multi-domain dialogue data to enhance the adaptability of our pre-trained model.
1 code implementation • 6 Jul 2023 • Min Xiao, Junnan Zhu, Haitao Lin, Yu Zhou, Chengqing Zong
Therefore, we propose a novel Coarse-to-Fine contribution network for multimodal Summarization (CFSum) to consider different contributions of images for summarization.
no code implementations • 6 Dec 2022 • Yang Zhao, Junnan Zhu, Lu Xiang, Jiajun Zhang, Yu Zhou, FeiFei Zhai, Chengqing Zong
To alleviate the CF, we investigate knowledge distillation based life-long learning methods.
2 code implementations • ACL 2022 • Haitao Lin, Junnan Zhu, Lu Xiang, Yu Zhou, Jiajun Zhang, Chengqing Zong
Therefore, we propose a novel role interaction enhanced method for role-oriented dialogue summarization.
2 code implementations • EMNLP 2021 • Haitao Lin, Liqun Ma, Junnan Zhu, Lu Xiang, Yu Zhou, Jiajun Zhang, Chengqing Zong
Therefore, in this paper, we introduce a novel Chinese dataset for Customer Service Dialogue Summarization (CSDS).
4 code implementations • ICLR 2021 • Aojun Zhou, Yukun Ma, Junnan Zhu, Jianbo Liu, Zhijie Zhang, Kun Yuan, Wenxiu Sun, Hongsheng Li
In this paper, we are the first to study training from scratch an N:M fine-grained structured sparse network, which can maintain the advantages of both unstructured fine-grained sparsity and structured coarse-grained sparsity simultaneously on specifically designed GPUs.
no code implementations • COLING 2020 • Yang Zhao, Lu Xiang, Junnan Zhu, Jiajun Zhang, Yu Zhou, Chengqing Zong
Previous studies combining knowledge graph (KG) with neural machine translation (NMT) have two problems: i) Knowledge under-utilization: they only focus on the entities that appear in both KG and training sentence pairs, making much knowledge in KG unable to be fully utilized.
no code implementations • COLING 2020 • Haoran Li, Junnan Zhu, Jiajun Zhang, Xiaodong He, Chengqing Zong
Thus, we propose a multimodal selective gate network that considers reciprocal relationships between textual and multi-level visual features, including global image descriptor, activation grids, and object proposals, to select highlights of the event when encoding the source sentence.
no code implementations • 28 Oct 2020 • Yuchen Liu, Junnan Zhu, Jiajun Zhang, Chengqing Zong
End-to-end speech translation aims to translate speech in one language into text in another language via an end-to-end way.
no code implementations • ACL 2020 • Junnan Zhu, Yu Zhou, Jiajun Zhang, Cheng-qing Zong
Cross-lingual summarization aims at summarizing a document in one language (e. g., Chinese) into another language (e. g., English).
1 code implementation • IJCNLP 2019 • Junnan Zhu, Qian Wang, Yining Wang, Yu Zhou, Jiajun Zhang, Shaonan Wang, Cheng-qing Zong
Moreover, we propose to further improve NCLS by incorporating two related tasks, monolingual summarization and machine translation, into the training process of CLS under multi-task learning.
no code implementations • EMNLP 2018 • Junnan Zhu, Haoran Li, Tianshang Liu, Yu Zhou, Jiajun Zhang, Cheng-qing Zong
In this paper, we propose a novel task, multimodal summarization with multimodal output (MSMO).
no code implementations • COLING 2018 • Haoran Li, Junnan Zhu, Jiajun Zhang, Cheng-qing Zong
In this paper, we investigate the sentence summarization task that produces a summary from a source sentence.
Ranked #7 on Text Summarization on DUC 2004 Task 1
no code implementations • EMNLP 2017 • Haoran Li, Junnan Zhu, Cong Ma, Jiajun Zhang, Cheng-qing Zong
In this work, we propose an extractive Multi-modal Summarization (MMS) method which can automatically generate a textual summary given a set of documents, images, audios and videos related to a specific topic.
Automatic Speech Recognition (ASR) Document Summarization +1