Search Results for author: Shijia Huang

Found 8 papers, 7 papers with code

LLaVA-Grounding: Grounded Visual Chat with Large Multimodal Models

1 code implementation5 Dec 2023 Hao Zhang, Hongyang Li, Feng Li, Tianhe Ren, Xueyan Zou, Shilong Liu, Shijia Huang, Jianfeng Gao, Lei Zhang, Chunyuan Li, Jianwei Yang

To address this issue, we have created GVC data that allows for the combination of grounding and chat capabilities.

CLEVA: Chinese Language Models EVAluation Platform

1 code implementation9 Aug 2023 Yanyang Li, Jianqiao Zhao, Duo Zheng, Zi-Yuan Hu, Zhi Chen, Xiaohui Su, Yongfeng Huang, Shijia Huang, Dahua Lin, Michael R. Lyu, LiWei Wang

With the continuous emergence of Chinese Large Language Models (LLMs), how to evaluate a model's capabilities has become an increasingly significant issue.

DQ-DETR: Dual Query Detection Transformer for Phrase Extraction and Grounding

1 code implementation28 Nov 2022 Shilong Liu, Yaoyuan Liang, Feng Li, Shijia Huang, Hao Zhang, Hang Su, Jun Zhu, Lei Zhang

As phrase extraction can be regarded as a $1$D text segmentation problem, we formulate PEG as a dual detection problem and propose a novel DQ-DETR model, which introduces dual queries to probe different features from image and text for object prediction and phrase mask prediction.

object-detection Object Detection +4

DSGN++: Exploiting Visual-Spatial Relation for Stereo-based 3D Detectors

1 code implementation6 Apr 2022 Yilun Chen, Shijia Huang, Shu Liu, Bei Yu, Jiaya Jia

First, to effectively lift the 2D information to stereo volume, we propose depth-wise plane sweeping (DPS) that allows denser connections and extracts depth-guided features.

3D Object Detection From Stereo Images Relation

Multi-View Transformer for 3D Visual Grounding

1 code implementation CVPR 2022 Shijia Huang, Yilun Chen, Jiaya Jia, LiWei Wang

The multi-view space enables the network to learn a more robust multi-modal representation for 3D visual grounding and eliminates the dependence on specific views.

Visual Grounding

Cannot find the paper you are looking for? You can Submit a new open access paper.