Search Results for author: Pengguang Chen

Found 20 papers, 15 papers with code

MTMamba++: Enhancing Multi-Task Dense Scene Understanding via Mamba-Based Decoders

1 code implementation27 Aug 2024 Baijiong Lin, Weisen Jiang, Pengguang Chen, Shu Liu, Ying-Cong Chen

Multi-task dense scene understanding, which trains a model for multiple dense prediction tasks, has a wide range of application scenarios.

Decoder Mamba +2

MTMamba: Enhancing Multi-Task Dense Scene Understanding by Mamba-Based Decoders

1 code implementation2 Jul 2024 Baijiong Lin, Weisen Jiang, Pengguang Chen, Yu Zhang, Shu Liu, Ying-Cong Chen

Multi-task dense scene understanding, which learns a model for multiple dense prediction tasks, has a wide range of application scenarios.

Boundary Detection Human Parsing +3

MR-Ben: A Meta-Reasoning Benchmark for Evaluating System-2 Thinking in LLMs

no code implementations20 Jun 2024 Zhongshen Zeng, Yinhong Liu, Yingjia Wan, Jingyao Li, Pengguang Chen, Jianbo Dai, Yuxuan Yao, Rongwu Xu, Zehan Qi, Wanru Zhao, Linling Shen, Jianqiao Lu, Haochen Tan, Yukang Chen, Hao Zhang, Zhan Shi, Bailin Wang, Zhijiang Guo, Jiaya Jia

Large language models (LLMs) have shown increasing capability in problem-solving and decision-making, largely based on the step-by-step chain-of-thought reasoning processes.

Decision Making

RoboCoder: Robotic Learning from Basic Skills to General Tasks with Large Language Models

no code implementations6 Jun 2024 Jingyao Li, Pengguang Chen, Sitong Wu, Chuanyang Zheng, Hong Xu, Jiaya Jia

To address these limitations, the RoboCoder framework integrates Large Language Models (LLMs) with a dynamic learning system that uses real-time environmental feedback to continuously update and refine action codes.

VLPose: Bridging the Domain Gap in Pose Estimation with Language-Vision Tuning

no code implementations22 Feb 2024 Jingyao Li, Pengguang Chen, Xuan Ju, Hong Xu, Jiaya Jia

Our research aims to bridge the domain gap between natural and artificial scenarios with efficient tuning strategies.

Pose Estimation

MOODv2: Masked Image Modeling for Out-of-Distribution Detection

1 code implementation5 Jan 2024 Jingyao Li, Pengguang Chen, Shaozuo Yu, Shu Liu, Jiaya Jia

The crux of effective out-of-distribution (OOD) detection lies in acquiring a robust in-distribution (ID) representation, distinct from OOD samples.

Out-of-Distribution Detection Out of Distribution (OOD) Detection

MR-GSM8K: A Meta-Reasoning Benchmark for Large Language Model Evaluation

2 code implementations28 Dec 2023 Zhongshen Zeng, Pengguang Chen, Shu Liu, Haiyun Jiang, Jiaya Jia

In this work, we introduce a novel evaluation paradigm for Large Language Models (LLMs) that compels them to transition from a traditional question-answering role, akin to a student, to a solution-scoring role, akin to a teacher.

GSM8K Language Modeling +4

BAL: Balancing Diversity and Novelty for Active Learning

1 code implementation26 Dec 2023 Jingyao Li, Pengguang Chen, Shaozuo Yu, Shu Liu, Jiaya Jia

Experimental results demonstrate that, when labeling 80% of the samples, the performance of the current SOTA method declines by 0. 74%, whereas our proposed BAL achieves performance comparable to the full dataset.

Active Learning Diversity +1

MoTCoder: Elevating Large Language Models with Modular of Thought for Challenging Programming Tasks

1 code implementation26 Dec 2023 Jingyao Li, Pengguang Chen, Bin Xia, Hong Xu, Jiaya Jia

Large Language Models (LLMs) have showcased impressive capabilities in handling straightforward programming tasks.

Code Generation

Dual-Balancing for Multi-Task Learning

1 code implementation23 Aug 2023 Baijiong Lin, Weisen Jiang, Feiyang Ye, Yu Zhang, Pengguang Chen, Ying-Cong Chen, Shu Liu, James T. Kwok

Multi-task learning (MTL), a learning paradigm to learn multiple related tasks simultaneously, has achieved great success in various fields.

Multi-Task Learning

TagCLIP: Improving Discrimination Ability of Open-Vocabulary Semantic Segmentation

1 code implementation15 Apr 2023 Jingyao Li, Pengguang Chen, Shengju Qian, Shu Liu, Jiaya Jia

Contrastive Language-Image Pre-training (CLIP) has recently shown great promise in pixel-level zero-shot learning tasks.

Language Modeling Language Modelling +5

Rethinking Out-of-distribution (OOD) Detection: Masked Image Modeling is All You Need

2 code implementations CVPR 2023 Jingyao Li, Pengguang Chen, Shaozuo Yu, Zexin He, Shu Liu, Jiaya Jia

The core of out-of-distribution (OOD) detection is to learn the in-distribution (ID) representation, which is distinguishable from OOD samples.

Out-of-Distribution Detection

SEA: Bridging the Gap Between One- and Two-stage Detector Distillation via SEmantic-aware Alignment

no code implementations2 Mar 2022 Yixin Chen, Zhuotao Tian, Pengguang Chen, Shu Liu, Jiaya Jia

We revisit the one- and two-stage detector distillation tasks and present a simple and efficient semantic-aware framework to fill the gap between them.

Instance Segmentation object-detection +2

Deep Structured Instance Graph for Distilling Object Detectors

1 code implementation ICCV 2021 Yixin Chen, Pengguang Chen, Shu Liu, LiWei Wang, Jiaya Jia

Effectively structuring deep knowledge plays a pivotal role in transfer from teacher to student, especially in semantic vision tasks.

Instance Segmentation Knowledge Distillation +5

Exploring and Improving Mobile Level Vision Transformers

no code implementations30 Aug 2021 Pengguang Chen, Yixin Chen, Shu Liu, MingChang Yang, Jiaya Jia

We analyze the reason behind this phenomenon, and propose a novel irregular patch embedding module and adaptive patch fusion module to improve the performance.

Distilling Knowledge via Knowledge Review

7 code implementations CVPR 2021 Pengguang Chen, Shu Liu, Hengshuang Zhao, Jiaya Jia

Knowledge distillation transfers knowledge from the teacher network to the student one, with the goal of greatly improving the performance of the student network.

Instance Segmentation Knowledge Distillation +3

GridMask Data Augmentation

7 code implementations13 Jan 2020 Pengguang Chen, Shu Liu, Hengshuang Zhao, Xingquan Wang, Jiaya Jia

Then we show limitation of existing information dropping algorithms and propose our structured method, which is simple and yet very effective.

Data Augmentation object-detection +5

Cannot find the paper you are looking for? You can Submit a new open access paper.