Search Results for author: Yu-Jung Heo

Found 12 papers, 3 papers with code

Scene Graph Parsing via Abstract Meaning Representation in Pre-trained Language Models

no code implementations NAACL (DLG4NLP) 2022 Woo Suk Choi, Yu-Jung Heo, Dharani Punithan, Byoung-Tak Zhang

In this work, we propose the application of abstract meaning representation (AMR) based semantic parsing models to parse textual descriptions of a visual scene into scene graphs, which is the first work to the best of our knowledge.

AMR Parsing Dependency Parsing

SGRAM: Improving Scene Graph Parsing via Abstract Meaning Representation

no code implementations17 Oct 2022 Woo Suk Choi, Yu-Jung Heo, Byoung-Tak Zhang

To this end, we design a simple yet effective two-stage scene graph parsing framework utilizing abstract meaning representation, SGRAM (Scene GRaph parsing via Abstract Meaning representation): 1) transforming a textual description of an image into an AMR graph (Text-to-AMR) and 2) encoding the AMR graph into a Transformer-based language model to generate a scene graph (AMR-to-SG).

Dependency Parsing Graph Generation +5

Hypergraph Transformer: Weakly-supervised Multi-hop Reasoning for Knowledge-based Visual Question Answering

1 code implementation ACL 2022 Yu-Jung Heo, Eun-Sol Kim, Woo Suk Choi, Byoung-Tak Zhang

Knowledge-based visual question answering (QA) aims to answer a question which requires visually-grounded external knowledge beyond image content itself.

Question Answering Visual Question Answering

Toward a Human-Level Video Understanding Intelligence

no code implementations8 Oct 2021 Yu-Jung Heo, Minsu Lee, SeongHo Choi, Woo Suk Choi, Minjung Shin, Minjoon Jung, Jeh-Kwang Ryu, Byoung-Tak Zhang

In this paper, we propose the Video Turing Test to provide effective and practical assessments of video understanding intelligence as well as human-likeness evaluation of AI agents.

Video Understanding

CogME: A Cognition-Inspired Multi-Dimensional Evaluation Metric for Story Understanding

no code implementations21 Jul 2021 Minjung Shin, SeongHo Choi, Yu-Jung Heo, Minsu Lee, Byoung-Tak Zhang, Jeh-Kwang Ryu

We introduce CogME, a cognition-inspired, multi-dimensional evaluation metric designed for AI models focusing on story understanding.

Question Answering Sentence +2

DramaQA: Character-Centered Video Story Understanding with Hierarchical QA

1 code implementation7 May 2020 Seong-Ho Choi, Kyoung-Woon On, Yu-Jung Heo, Ahjeong Seo, Youwon Jang, Minsu Lee, Byoung-Tak Zhang

Despite recent progress on computer vision and natural language processing, developing a machine that can understand video story is still hard to achieve due to the intrinsic difficulty of video story.

Question Answering Video Question Answering +1

Cut-Based Graph Learning Networks to Discover Compositional Structure of Sequential Video Data

no code implementations17 Jan 2020 Kyoung-Woon On, Eun-Sol Kim, Yu-Jung Heo, Byoung-Tak Zhang

Here, we propose Cut-Based Graph Learning Networks (CB-GLNs) for learning video data by discovering these complex structures of the video.

Graph Learning Video Understanding

Compositional Structure Learning for Sequential Video Data

no code implementations3 Jul 2019 Kyoung-Woon On, Eun-Sol Kim, Yu-Jung Heo, Byoung-Tak Zhang

However, most of sequential data, as seen with videos, have complex temporal dependencies that imply variable-length semantic flows and their compositions, and those are hard to be captured by conventional methods.

Visualizing Semantic Structures of Sequential Data by Learning Temporal Dependencies

no code implementations20 Jan 2019 Kyoung-Woon On, Eun-Sol Kim, Yu-Jung Heo, Byoung-Tak Zhang

While conventional methods for sequential learning focus on interaction between consecutive inputs, we suggest a new method which captures composite semantic flows with variable-length dependencies.

Answerer in Questioner's Mind: Information Theoretic Approach to Goal-Oriented Visual Dialog

1 code implementation NeurIPS 2018 Sang-Woo Lee, Yu-Jung Heo, Byoung-Tak Zhang

Goal-oriented dialogue tasks occur when a questioner asks an action-oriented question and an answerer responds with the intent of letting the questioner know a correct action to take.

Goal-Oriented Dialog Visual Dialog

Cannot find the paper you are looking for? You can Submit a new open access paper.