no code implementations • 30 Apr 2024 • Zhihao Zhang, Feiqi Cao, Yingbin Mo, Yiran Zhang, Josiah Poon, Caren Han
In addition, we also propose a new audience conversation augmented commentary dataset by covering the game situation and audience conversation understanding, and introducing a robust joint multimodal dual learning model as a baseline.
1 code implementation • 21 Apr 2024 • Feiqi Cao, Caren Han, Hyunsuk Chung
In this work, we propose a novel tree-based explanation technique, PEACH (Pretrained-embedding Explanation Across Contextual and Hierarchical Structure), that can explain how text-based documents are classified by using any pretrained contextual embeddings in a tree-based human-interpretable manner.
no code implementations • 16 Dec 2022 • Feiqi Cao, Siwen Luo, Felipe Nunez, Zean Wen, Josiah Poon, Caren Han
To make explicit teaching of the relations between the two modalities, we proposed and integrated two attention modules, namely a scene graph-based semantic relation-aware attention and a positional relation-aware attention.
Optical Character Recognition Optical Character Recognition (OCR) +3
no code implementations • 11 Nov 2022 • Yuanzhe Jia, Weixuan Wu, Feiqi Cao, Soyeon Caren Han
In-game toxic language becomes the hot potato in the gaming industry and community.
1 code implementation • COLING 2022 • Feiqi Cao, Soyeon Caren Han, Siqu Long, Changwei Xu, Josiah Poon
Attention mechanism has been used as an important component across Vision-and-Language(VL) tasks in order to bridge the semantic gap between visual and textual features.
no code implementations • 15 Apr 2022 • Siqu Long, Feiqi Cao, Soyeon Caren Han, Haiqin Yang
Pretrained models have produced great success in both Computer Vision (CV) and Natural Language Processing (NLP).