Search Results for author: Yifan Hao

Found 22 papers, 5 papers with code

Towards Better Generalization via Distributional Input Projection Network

no code implementations5 Jun 2025 Yifan Hao, Yanxin Lu, Xinwei Shen, Tong Zhang

As overparameterized models become increasingly prevalent, training loss alone offers limited insight into generalization performance.

QiMeng: Fully Automated Hardware and Software Design for Processor Chip

no code implementations5 Jun 2025 Rui Zhang, Yuanbo Wen, Shuyao Cheng, Di Huang, Shaohui Peng, Jiaming Guo, Pengwei Jin, Jiacheng Zhao, Tianrui Ma, Yaoyu Zhu, Yifan Hao, Yongwei Zhao, Shengwen Liang, Ying Wang, Xing Hu, Zidong Du, Huimin Cui, Ling Li, Qi Guo, Yunji Chen

In the middle-layer, leveraging the LPCM's knowledge representation and inference capabilities, we develop the Hardware Design Agent and the Software Design Agent to automate the design of hardware and software for processor chips.

Optimizing Chain-of-Thought Reasoners via Gradient Variance Minimization in Rejection Sampling and RL

1 code implementation5 May 2025 Jiarui Yao, Yifan Hao, Hanning Zhang, Hanze Dong, Wei Xiong, Nan Jiang, Tong Zhang

Chain-of-thought (CoT) reasoning in large language models (LLMs) can be formalized as a latent variable problem, where the model needs to generate intermediate reasoning steps.

Mathematical Reasoning

QiMeng-Xpiler: Transcompiling Tensor Programs for Deep Learning Systems with a Neural-Symbolic Approach

no code implementations4 May 2025 Shouyang Dong, Yuanbo Wen, Jun Bi, Di Huang, Jiaming Guo, Jianxing Xu, Ruibai Xu, Xinkai Song, Yifan Hao, Xuehai Zhou, Tianshi Chen, Qi Guo, Yunji Chen

We propose a novel transcompiler, i. e., QiMeng-Xpiler, for automatically translating tensor programs across DLS via both large language models (LLMs) and symbolic program synthesis, i. e., neural-symbolic synthesis.

Code Generation Program Synthesis

Object-Level Verbalized Confidence Calibration in Vision-Language Models via Semantic Perturbation

no code implementations21 Apr 2025 Yunpu Zhao, Rui Zhang, Junbin Xiao, Ruibo Hou, Jiaming Guo, Zihao Zhang, Yifan Hao, Yunji Chen

In this work, we propose a novel Confidence Calibration through Semantic Perturbation (CSP) framework to improve the calibration of verbalized confidence for VLMs in response to object-centric queries.

Efficient Model Editing with Task Vector Bases: A Theoretical Framework and Scalable Approach

1 code implementation3 Feb 2025 Siqi Zeng, Yifei He, Weiqiu You, Yifan Hao, Yao-Hung Hubert Tsai, Makoto Yamada, Han Zhao

Task vectors, which are derived from the difference between pre-trained and fine-tuned model weights, enable flexible task adaptation and model merging through arithmetic operations such as addition and negation.

Model Editing Negation +1

DA-Ada: Learning Domain-Aware Adapter for Domain Adaptive Object Detection

1 code implementation11 Oct 2024 Haochen Li, Rui Zhang, Hantao Yao, Xin Zhang, Yifan Hao, Xinkai Song, Xiaqing Li, Yongwei Zhao, Ling Li, Yunji Chen

Domain adaptive object detection (DAOD) aims to generalize detectors trained on an annotated source domain to an unlabelled target domain.

General Knowledge object-detection +1

SPDiffusion: Semantic Protection Diffusion for Multi-concept Text-to-image Generation

no code implementations2 Sep 2024 Yang Zhang, Rui Zhang, Xuecheng Nie, Haochen Li, Jikun Chen, Yifan Hao, Xin Zhang, Luoqi Liu, Ling Li

We found that attribute confusion occurs when a certain region of the latent features attend to multiple or incorrect prompt tokens.

Attribute Text to Image Generation +1

Towards Analyzing and Mitigating Sycophancy in Large Vision-Language Models

no code implementations21 Aug 2024 Yunpu Zhao, Rui Zhang, Junbin Xiao, Changxin Ke, Ruibo Hou, Yifan Hao, Qi Guo, Yunji Chen

For improvement, we propose Leading Query Contrastive Decoding (LQCD), a model-agnostic method focusing on calibrating the LVLMs' over-reliance on leading cues by identifying and suppressing the probabilities of sycophancy tokens at the decoding stage.

Hallucination Prompt Engineering

HeteroMorpheus: Universal Control Based on Morphological Heterogeneity Modeling

1 code implementation2 Aug 2024 Yifan Hao, Yang Yang, Junru Song, Wei Peng, Weien Zhou, Tingsong Jiang, Wen Yao

In the field of robotic control, designing individual controllers for each robot leads to high computational costs.

Diversity Zero-shot Generalization

On the Benefits of Over-parameterization for Out-of-Distribution Generalization

no code implementations26 Mar 2024 Yifan Hao, Yong Lin, Difan Zou, Tong Zhang

We demonstrate that in this scenario, further increasing the model's parameterization can significantly reduce the OOD loss.

Out-of-Distribution Generalization

The Surprising Harmfulness of Benign Overfitting for Adversarial Robustness

no code implementations19 Jan 2024 Yifan Hao, Tong Zhang

Recent empirical and theoretical studies have established the generalization capabilities of large machine learning models that are trained to (approximately or exactly) fit noisy data.

Adversarial Robustness

Spurious Feature Diversification Improves Out-of-distribution Generalization

no code implementations29 Sep 2023 Yong Lin, Lu Tan, Yifan Hao, Honam Wong, Hanze Dong, Weizhong Zhang, Yujiu Yang, Tong Zhang

Contrary to the conventional wisdom that focuses on learning invariant features for better OOD performance, our findings suggest that incorporating a large number of diverse spurious features weakens their individual contributions, leading to improved overall OOD generalization performance.

Out-of-Distribution Generalization

Reverse Diffusion Monte Carlo

no code implementations5 Jul 2023 Xunpeng Huang, Hanze Dong, Yifan Hao, Yi-An Ma, Tong Zhang

We propose a Monte Carlo sampler from the reverse diffusion process.

Pushing the Limits of Machine Design: Automated CPU Design with AI

1 code implementation21 Jun 2023 Shuyao Cheng, Pengwei Jin, Qi Guo, Zidong Du, Rui Zhang, Yunhao Tian, Xing Hu, Yongwei Zhao, Yifan Hao, Xiangtao Guan, Husheng Han, Zhengyue Zhao, Ximing Liu, Ling Li, Xishan Zhang, Yuejie Chu, Weilong Mao, Tianshi Chen, Yunji Chen

By efficiently exploring a search space of unprecedented size 10^{10^{540}}, which is the largest one of all machine-designed objects to our best knowledge, and thus pushing the limits of machine design, our approach generates an industrial-scale RISC-V CPU within only 5 hours.

Ultra-low Precision Multiplication-free Training for Deep Neural Networks

no code implementations28 Feb 2023 Chang Liu, Rui Zhang, Xishan Zhang, Yifan Hao, Zidong Du, Xing Hu, Ling Li, Qi Guo

The energy-efficient works try to decrease the precision of multiplication or replace the multiplication with energy-efficient operations such as addition or bitwise shift, to reduce the energy consumption of FP32 multiplications.

Quantization

Class-Specific Attention (CSA) for Time-Series Classification

no code implementations19 Nov 2022 Yifan Hao, Huiping Cao, K. Selcuk Candan, Jiefei Liu, Huiying Chen, Ziwei Ma

In this paper, we propose a novel class-specific attention (CSA) module to capture significant class-specific features and improve the overall classification performance of time series.

Classification Time Series +2

Scheduling Real-time Deep Learning Services as Imprecise Computations

no code implementations2 Nov 2020 Shuochao Yao, Yifan Hao, Yiran Zhao, Huajie Shao, Dongxin Liu, Shengzhong Liu, Tianshi Wang, Jinyang Li, Tarek Abdelzaher

The paper presents an efficient real-time scheduling algorithm for intelligent real-time edge services, defined as those that perform machine intelligence tasks, such as voice recognition, LIDAR processing, or machine vision, on behalf of local embedded devices that are themselves unable to support extensive computations.

Deep Learning Scheduling

Cannot find the paper you are looking for? You can Submit a new open access paper.