Search Results for author: Yuan Meng

Found 32 papers, 10 papers with code

PhysUniBench: An Undergraduate-Level Physics Reasoning Benchmark for Multimodal Models

no code implementations21 Jun 2025 Lintao Wang, Encheng Su, Jiaqi Liu, Pengze Li, Peng Xia, Jiabei Xiao, Wenlong Zhang, Xinnan Dai, Xi Chen, Yuan Meng, Mingyu Ding, Lei Bai, Wanli Ouyang, Shixiang Tang, Aoran Wang, Xinzhu Ma

These results highlight that current MLLMs struggle with advanced physics reasoning, especially on multi-step problems and those requiring precise diagram interpretation.

Mathematical Reasoning Multiple-choice

Block-wise Adaptive Caching for Accelerating Diffusion Policy

no code implementations16 Jun 2025 Kangye Ji, Yuan Meng, Hanyun Cui, Ye Li, Shengjia Hua, Lei Chen, Zhi Wang

BAC achieves lossless action generation acceleration by adaptively updating and reusing cached features at the block level, based on a key observation that feature similarities vary non-uniformly across timesteps and locks.

Action Generation Denoising +1

Point2Primitive: CAD Reconstruction from Point Cloud by Direct Primitive Prediction

no code implementations4 May 2025 Cheng Wang, Xinzhu Ma, Bin Wang, Shixiang Tang, Yuan Meng, Ping Jiang

Recovering CAD models from point clouds, especially the sketch-extrusion process, can be seen as the process of rebuilding the topology and extrusion primitives.

CAD Reconstruction

An Efficient and Mixed Heterogeneous Model for Image Restoration

1 code implementation15 Apr 2025 Yubin Gu, Yuan Meng, Kaihang Zheng, Xiaoshuai Sun, Jiayi Ji, Weijian Ruan, Liujuan Cao, Rongrong Ji

This hierarchical and adaptive design enables the model to leverage the strengths of CNNs in local feature extraction, Mamba in global context modeling, and attention mechanisms in dynamic feature refinement.

Image Restoration Mamba

Pretrained Bayesian Non-parametric Knowledge Prior in Robotic Long-Horizon Reinforcement Learning

1 code implementation27 Mar 2025 Yuan Meng, Xiangtong Yao, KeJia Chen, Yansong Wu, Liding Zhang, Zhenshan Bing, Alois Knoll

Reinforcement learning (RL) methods typically learn new tasks from scratch, often disregarding prior knowledge that could accelerate the learning process.

Reinforcement Learning (RL)

Data-Agnostic Robotic Long-Horizon Manipulation with Vision-Language-Guided Closed-Loop Feedback

1 code implementation27 Mar 2025 Yuan Meng, Xiangtong Yao, Haihui Ye, Yirui Zhou, Shengqiang Zhang, Zhenshan Bing, Alois Knoll

Recent advances in language-conditioned robotic manipulation have leveraged imitation and reinforcement learning to enable robots to execute tasks from human commands.

Task Planning

3DAxisPrompt: Promoting the 3D Grounding and Reasoning in GPT-4o

no code implementations17 Mar 2025 Dingning Liu, Cheng Wang, Peng Gao, Renrui Zhang, Xinzhu Ma, Yuan Meng, Zhihui Wang

Multimodal Large Language Models (MLLMs) exhibit impressive capabilities across a variety of tasks, especially when equipped with carefully designed visual prompts.

Logical Reasoning Visual Prompting

JAQ: Joint Efficient Architecture Design and Low-Bit Quantization with Hardware-Software Co-Exploration

no code implementations9 Jan 2025 Mingzi Wang, Yuan Meng, Chen Tang, Weixiang Zhang, Yijian Qin, Yang Yao, Yingxin Li, Tongtong Feng, Xin Wang, Xun Guan, Zhi Wang, Wenwu Zhu

The co-design of neural network architectures, quantization precisions, and hardware accelerators offers a promising approach to achieving an optimal balance between performance and efficiency, particularly for model deployment on resource-constrained edge devices.

Quantization

ACL: Activating Capability of Linear Attention for Image Restoration

no code implementations CVPR 2025 Yubin Gu, Yuan Meng, Jiayi Ji, Xiaoshuai Sun

ACL integrates linear attention blocks instead of SSM within Mamba, serving as the core component of encoders/decoders, and aims to preserve a global perspective while boosting computational efficiency.

Computational Efficiency Image Restoration +2

EMS: Adaptive Evict-then-Merge Strategy for Head-wise KV Cache Compression Based on Global-Local Importance

no code implementations11 Dec 2024 Yingxin Li, Ye Li, Yuan Meng, Xinzhu Ma, Zihan Geng, Shutao Xia, Zhi Wang

However, these approaches often produce biased distributions of important tokens due to the influence of accumulated attention scores or positional encoding.

GAQAT: gradient-adaptive quantization-aware training for domain generalization

no code implementations7 Dec 2024 Jiacheng Jiang, Yuan Meng, Chen Tang, Han Yu, Qun Li, Zhi Wang, Wenwu Zhu

To address this limitation, we propose a novel Gradient-Adaptive Quantization-Aware Training (GAQAT) framework for DG.

Domain Generalization Quantization

CATP-LLM: Empowering Large Language Models for Cost-Aware Tool Planning

no code implementations25 Nov 2024 Duo Wu, Jinghe Wang, Yuan Meng, Yanning Zhang, Le Sun, Zhi Wang

To push this paradigm toward practical applications, it is crucial for LLMs to consider tool execution costs (e. g. execution time) for tool planning.

Mixed Degradation Image Restoration via Local Dynamic Optimization and Conditional Embedding

no code implementations25 Nov 2024 Yubin Gu, Yuan Meng, Xiaoshuai Sun, Jiayi Ji, Weijian Ruan, Rongrong Ji

In this paper, we propose a novel multiple-in-one IR model that can effectively restore images with both single and mixed degradations.

Decoder Diversity +1

RTF-Q: Efficient Unsupervised Domain Adaptation with Retraining-free Quantization

no code implementations11 Aug 2024 Nanyang Du, Chen Tang, Yuxiao Jiang, Yuan Meng, Zhi Wang

To address these limitations, we propose efficient unsupervised domain adaptation with ReTraining-Free Quantization (RTF-Q).

Quantization Unsupervised Domain Adaptation

PRANCE: Joint Token-Optimization and Structural Channel-Pruning for Adaptive ViT Inference

1 code implementation6 Jul 2024 Ye Li, Chen Tang, Yuan Meng, Jiajun Fan, Zenghao Chai, Xinzhu Ma, Zhi Wang, Wenwu Zhu

We introduce PRANCE, a Vision Transformer compression framework that jointly optimizes the activated channels and reduces tokens, based on the characteristics of inputs.

Combinatorial Optimization Decision Making

Q-DiT: Accurate Post-Training Quantization for Diffusion Transformers

1 code implementation CVPR 2025 Lei Chen, Yuan Meng, Chen Tang, Xinzhu Ma, Jingyan Jiang, Xin Wang, Zhi Wang, Wenwu Zhu

Recent advancements in diffusion models, particularly the architectural transformation from UNet-based models to Diffusion Transformers (DiTs), significantly improve the quality and scalability of image and video generation.

Image Generation Model Compression +2

Towards Lightweight Graph Neural Network Search with Curriculum Graph Sparsification

no code implementations24 Jun 2024 Beini Xie, Heng Chang, Ziwei Zhang, Zeyang Zhang, Simin Wu, Xin Wang, Yuan Meng, Wenwu Zhu

To search for optimal lightweight Graph Neural Networks (GNNs), we propose a Lightweight Graph Neural Architecture Search with Graph SparsIfication and Network Pruning (GASSIP) method.

Graph Neural Network Network Pruning +2

Evaluating the Generalization Ability of Quantized LLMs: Benchmark, Analysis, and Toolbox

1 code implementation15 Jun 2024 Yijun Liu, Yuan Meng, Fang Wu, Shenhao Peng, Hang Yao, Chaoyu Guan, Chen Tang, Xinzhu Ma, Zhi Wang, Wenwu Zhu

Based on this benchmark, we conduct extensive experiments with two well-known LLMs (English and Chinese) and four quantization algorithms to investigate this topic in-depth, yielding several counter-intuitive and valuable findings, e. g., models quantized using a calibration set with the same distribution as the test data are not necessarily optimal.

Quantization

RealTCD: Temporal Causal Discovery from Interventional Data with Large Language Model

no code implementations23 Apr 2024 Peiwen Li, Xin Wang, Zeyang Zhang, Yuan Meng, Fang Shen, Yue Li, Jialong Wang, Yang Li, Wenweu Zhu

In the field of Artificial Intelligence for Information Technology Operations, causal discovery is pivotal for operation and maintenance of graph construction, facilitating downstream industrial tasks such as root cause analysis.

Causal Discovery graph construction +3

Investigating the Impact of Quantization on Adversarial Robustness

no code implementations8 Apr 2024 Qun Li, Yuan Meng, Chen Tang, Jiacheng Jiang, Zhi Wang

Quantization is a promising technique for reducing the bit-width of deep models to improve their runtime performance and storage efficiency, and thus becomes a fundamental step for deployment.

Adversarial Robustness Quantization

Retraining-free Model Quantization via One-Shot Weight-Coupling Learning

1 code implementation CVPR 2024 Chen Tang, Yuan Meng, Jiacheng Jiang, Shuzhao Xie, Rongwei Lu, Xinzhu Ma, Zhi Wang, Wenwu Zhu

Conversely, mixed-precision quantization (MPQ) is advocated to compress the model effectively by allocating heterogeneous bit-width for layers.

Model Compression Quantization

Towards Fair and Comprehensive Comparisons for Image-Based 3D Object Detection

no code implementations ICCV 2023 Xinzhu Ma, Yongtao Wang, Yinmin Zhang, Zhiyi Xia, Yuan Meng, Zhihui Wang, Haojie Li, Wanli Ouyang

In this work, we build a modular-designed codebase, formulate strong training recipes, design an error diagnosis toolbox, and discuss current methods for image-based 3D object detection.

3D Object Detection Object +1

Characterizing Speed Performance of Multi-Agent Reinforcement Learning

no code implementations13 Sep 2023 Samuel Wiggins, Yuan Meng, Rajgopal Kannan, Viktor Prasanna

Multi-Agent Reinforcement Learning (MARL) has achieved significant success in large-scale AI systems and big-data applications such as smart grids, surveillance, etc.

Multi-agent Reinforcement Learning reinforcement-learning +1

RGAT: A Deeper Look into Syntactic Dependency Information for Coreference Resolution

no code implementations10 Sep 2023 Yuan Meng, Xuhao Pan, Jun Chang, Yue Wang

Our experiments on a public Gendered Ambiguous Pronouns (GAP) dataset show that with the supervision learning of the syntactic dependency graph and without fine-tuning the entire BERT, we increased the F1-score of the previous best model (RGCN-with-BERT) from 80. 3% to 82. 5%, compared to the F1-score by single BERT embeddings from 78. 5% to 82. 5%.

coreference-resolution Graph Attention

DIVA: A Dirichlet Process Mixtures Based Incremental Deep Clustering Algorithm via Variational Auto-Encoder

1 code implementation23 May 2023 Zhenshan Bing, Yuan Meng, Yuqi Yun, Hang Su, Xiaojie Su, Kai Huang, Alois Knoll

Generative model-based deep clustering frameworks excel in classifying complex data, but are limited in handling dynamic and complex features because they require prior knowledge of the number of clusters.

Clustering Image Generation +2

DNG: Taxonomy Expansion by Exploring the Intrinsic Directed Structure on Non-gaussian Space

1 code implementation22 Feb 2023 Songlin Zhai, Weiqing Wang, YuanFang Li, Yuan Meng

Specifically, the inherited feature originates from "parent" nodes and is weighted by an inheritance factor.

Taxonomy Expansion

SEAM: Searching Transferable Mixed-Precision Quantization Policy through Large Margin Regularization

no code implementations14 Feb 2023 Chen Tang, Kai Ouyang, Zenghao Chai, Yunpeng Bai, Yuan Meng, Zhi Wang, Wenwu Zhu

This general and dataset-independent property makes us search for the MPQ policy over a rather small-scale proxy dataset and then the policy can be directly used to quantize the model trained on a large-scale dataset.

Quantization

Estimating Granger Causality with Unobserved Confounders via Deep Latent-Variable Recurrent Neural Network

no code implementations9 Sep 2019 Yuan Meng

We use a generative model with latent variable to build the relationship between the unobserved confounders and the observed variables(tested variable and the proxy variables).

Time Series Analysis

Cannot find the paper you are looking for? You can Submit a new open access paper.