no code implementations • 9 Apr 2025 • Enming Zhang, Liwen Cao, Yanru Wu, Zijie Zhao, Guan Wang, Yang Li
Prompt tuning has emerged as a lightweight adaptation strategy for adapting foundation models to downstream tasks, particularly in resource-constrained systems.
no code implementations • 9 Apr 2025 • Enming Zhang, Zheng Liu, Yu Xiang, Yanwen Qu
Active QoS metric prediction, commonly employed in the maintenance and operation of DTN, could enhance network performance regarding latency, throughput, energy consumption, and dependability.
Multivariate Time Series Forecasting
Probabilistic Time Series Forecasting
+1
no code implementations • 9 Mar 2025 • Enming Zhang, Peizhe Gong, Xingyuan Dai, Yisheng Lv, Qinghai Miao
However, since lightweight models are crucial for autonomous driving systems, this presents a significant challenge for integrating VLMs into the field.
no code implementations • 17 Feb 2025 • Yanru Wu, Xiangyu Chen, Jianning Wang, Enming Zhang, Hanbing Liu, Yang Li
Continual learning (CL) has been an essential topic in the contemporary application of deep neural networks, where catastrophic forgetting (CF) can impede a model's ability to acquire knowledge progressively.
1 code implementation • 11 Sep 2024 • Enming Zhang, Xingyuan Dai, Min Huang, Yisheng Lv, Qinghai Miao
Meanwhile, most existing VLMs lack the ability to process multiple images, making it difficult to adapt to multi-camera perception in autonomous driving.
2 code implementations • 14 Jun 2024 • Enming Zhang, Ruobing Yao, Huanyong Liu, Junhui Yu, Jiale Wang
We propose the first comprehensive method, FlowCE, to assess MLLMs across various dimensions for tasks related to flowcharts.
no code implementations • 16 Apr 2024 • Enming Zhang, Bingke Zhu, Yingying Chen, Qinghai Miao, Ming Tang, Jinqiao Wang
This limitation restricts the capabilities of pretrained VLMs and can result in incorrect predictions in downstream tasks.
1 code implementation • 21 Mar 2024 • Zheng Zhang, Yeyao Ma, Enming Zhang, Xiang Bai
PSALM is a powerful extension of the Large Multi-modal Model (LMM) to address the segmentation task challenges.
Ranked #2 on
Referring Expression Segmentation
on RefCoCo val
(using extra training data)
1 code implementation • 14 Dec 2023 • Shuailei Ma, Yuefeng Wang, Ying WEI, Jiaqi Fan, Enming Zhang, Xinyu Sun, Peihao Chen
Ablation experiments demonstrate that both of them are effective in mitigating the impact of open-world knowledge distillation on the learning of known objects.
2 code implementations • 15 Nov 2023 • Junqing He, Kunhao Pan, Xiaoqun Dong, Zhuoyang Song, Yibo Liu, Qianguo Sun, Yuxin Liang, Hao Wang, Enming Zhang, Jiaxing Zhang
While large language models (LLMs) are equipped with longer text input capabilities than before, they are struggling to seek correct information in long contexts.
1 code implementation • 6 Jun 2023 • Wenwen Yu, MingYu Liu, Biao Yang, Enming Zhang, Deqiang Jiang, Xing Sun, Yuliang Liu, Xiang Bai
Text recognition in the wild is a long-standing problem in computer vision.
1 code implementation • 21 Mar 2023 • Shuailei Ma, Yuefeng Wang, Ying WEI, Peihao Chen, Zhixiang Ye, Jiaqi Fan, Enming Zhang, Thomas H. Li
We propose leveraging the VL as the ``Brain'' of the open-world detector by simply generating unknown labels.
no code implementations • 12 Jul 2022 • Yang Tan, Enming Zhang, Yang Li, Shao-Lun Huang, Xiao-Ping Zhang
We propose two novel transferability metrics F-OTCE (Fast Optimal Transport based Conditional Entropy) and JC-OTCE (Joint Correspondence OTCE) to evaluate how much the source model (task) can benefit the learning of the target task and to learn more transferable representations for cross-domain cross-task transfer learning.