Search Results for author: Yijiang Li

Found 26 papers, 7 papers with code

VideoOrion: Tokenizing Object Dynamics in Videos

no code implementations25 Nov 2024 Yicheng Feng, Yijiang Li, Wanpeng Zhang, Sipeng Zheng, Zongqing Lu

We present VideoOrion, a Video Large Language Model (Video-LLM) that explicitly captures the key semantic information in videos--the spatial-temporal dynamics of objects throughout the videos.

Language Modeling Language Modelling +4

From RNNs to Foundation Models: An Empirical Study on Commercial Building Energy Consumption

no code implementations21 Nov 2024 Shourya Bose, Yijiang Li, Amy Van Sant, Yu Zhang, Kibaek Kim

The impact of increasing dataset heterogeneity in time series forecasting, while keeping size and model constant, is understudied.

Diversity Time Series +1

Pseudo-Probability Unlearning: Towards Efficient and Privacy-Preserving Machine Unlearning

no code implementations4 Nov 2024 Zihao Zhao, Yijiang Li, Yuchen Yang, Wenqing Zhang, Nuno Vasconcelos, Yinzhi Cao

Machine unlearning--enabling a trained model to forget specific data--is crucial for addressing biased data and adhering to privacy regulations like the General Data Protection Regulation (GDPR)'s "right to be forgotten".

Machine Unlearning Privacy Preserving

FedDTPT: Federated Discrete and Transferable Prompt Tuning for Black-Box Large Language Models

no code implementations1 Nov 2024 Jiaqi Wu, Simin Chen, Yuzhe Yang, Yijiang Li, Shiyue Hou, Rui Jing, Zehua Wang, Wei Chen, Zijian Tian

To address these challenges, we propose for the first time a federated discrete and transferable prompt tuning, namely FedDTPT, for black-box large language models.

Federated Learning Semantic Similarity +1

FedSpaLLM: Federated Pruning of Large Language Models

no code implementations18 Oct 2024 Guangji Bai, Yijiang Li, Zilinghan Li, Liang Zhao, Kibaek Kim

Large Language Models (LLMs) achieve state-of-the-art performance but are challenging to deploy due to their high computational and storage demands.

Federated Learning Privacy Preserving

CogDevelop2K: Reversed Cognitive Development in Multimodal Large Language Models

no code implementations6 Oct 2024 Yijiang Li, Qingying Gao, Haoran Sun, Haiyun Lyu, Dezhi Luo, Hokin Deng

To this end, we propose CogDevelop2K, a comprehensive benchmark that spans 12 sub-concepts from primitive knowledge like object permanence and boundary to more complex abilities like intentionality understanding, structured via the developmental trajectory of a human mind.

From Pixels to Tokens: Byte-Pair Encoding on Quantized Visual Modalities

no code implementations3 Oct 2024 Wanpeng Zhang, Zilong Xie, Yicheng Feng, Yijiang Li, Xingrun Xing, Sipeng Zheng, Zongqing Lu

Multimodal Large Language Models have made significant strides in integrating visual and textual information, yet they often struggle with effectively aligning these modalities.

Vision Language Models Know Law of Conservation without Understanding More-or-Less

no code implementations1 Oct 2024 Dezhi Luo, Haiyun Lyu, Qingying Gao, Haoran Sun, Yijiang Li, Hokin Deng

Conservation is a critical milestone of cognitive development considered to be supported by both the understanding of quantitative concepts and the reversibility of mental operations.

Vision Language Models See What You Want but not What You See

no code implementations1 Oct 2024 Qingying Gao, Yijiang Li, Haiyun Lyu, Haoran Sun, Dezhi Luo, Hokin Deng

Knowing others' intentions and taking others' perspectives are two core components of human intelligence typically considered as instantiations of theory of mind.

Probing Mechanical Reasoning in Large Vision Language Models

no code implementations1 Oct 2024 Haoran Sun, Qingying Gao, Haiyun Lyu, Dezhi Luo, Hokin Deng, Yijiang Li

Mechanical reasoning is a fundamental ability that sets human intelligence apart from other animal intelligence.

ARVideo: Autoregressive Pretraining for Self-Supervised Video Representation Learning

no code implementations24 May 2024 Sucheng Ren, Hongru Zhu, Chen Wei, Yijiang Li, Alan Yuille, Cihang Xie

This paper presents a new self-supervised video representation learning framework, ARVideo, which autoregressively predicts the next video token in a tailored sequence order.

Representation Learning

Jailbreaking Prompt Attack: A Controllable Adversarial Attack against Diffusion Models

no code implementations2 Apr 2024 Jiachen Ma, Anda Cao, Zhiqing Xiao, Yijiang Li, Jie Zhang, Chao Ye, Junbo Zhao

In this work, we investigate a more practical and universal attack that does not require the presence of a target model and demonstrate that the high-dimensional text embedding space inherently contains NSFW concepts that can be exploited to generate harmful images.

Adversarial Attack Text-to-Image Generation

Can 3D Vision-Language Models Truly Understand Natural Language?

1 code implementation21 Mar 2024 Weipeng Deng, Jihan Yang, Runyu Ding, Jiahui Liu, Yijiang Li, Xiaojuan Qi, Edith Ngai

To test the language understandability of 3D-VL models, we first propose a language robustness task for systematically assessing 3D-VL models across various tasks, benchmarking their performance when presented with different language style variants.

Benchmarking Diversity

Towards Adversarially Robust Dataset Distillation by Curvature Regularization

no code implementations15 Mar 2024 Eric Xue, Yijiang Li, Haoyang Liu, Peiran Wang, Yifan Shen, Haohan Wang

Extensive empirical experiments suggest that our method not only outperforms standard adversarial training on both accuracy and robustness with less computation overhead but is also capable of generating robust distilled datasets that can withstand various adversarial attacks.

Adversarial Robustness Dataset Distillation

Approximate Nullspace Augmented Finetuning for Robust Vision Transformers

no code implementations15 Mar 2024 Haoyang Liu, Aditya Singh, Yijiang Li, Haohan Wang

In this work, we provide a finetuning approach to enhance the robustness of vision transformers inspired by the concept of nullspace from linear algebra.

SparseLLM: Towards Global Pruning for Pre-trained Language Models

2 code implementations28 Feb 2024 Guangji Bai, Yijiang Li, Chen Ling, Kibaek Kim, Liang Zhao

The transformative impact of large language models (LLMs) like LLaMA and GPT on natural language processing is countered by their prohibitive computational demands.

Computational Efficiency Problem Decomposition

Dataset Distillation via the Wasserstein Metric

no code implementations30 Nov 2023 Haoyang Liu, Yijiang Li, Tiancheng Xing, Vibhu Dalal, Luwei Li, Jingrui He, Haohan Wang

Dataset Distillation (DD) emerges as a powerful strategy to encapsulate the expansive information of large datasets into significantly smaller, synthetic equivalents, thereby preserving model performance with reduced computational overhead.

Dataset Distillation

Choosing Wisely and Learning Deeply: Selective Cross-Modality Distillation via CLIP for Domain Generalization

1 code implementation26 Nov 2023 Jixuan Leng, Yijiang Li, Haohan Wang

SCMD leverages the capabilities of large vision-language models, specifically CLIP, to train a more efficient model, ensuring it acquires robust generalization capabilities across unseen domains.

Domain Generalization

Towards Understanding Adversarial Transferability in Federated Learning

no code implementations1 Oct 2023 Yijiang Li, Ying Gao, Haohan Wang

We investigate a specific security risk in FL: a group of malicious clients has impacted the model during training by disguising their identities and acting as benign clients but later switching to an adversarial role.

Attribute Federated Learning

Diverse Cotraining Makes Strong Semi-Supervised Segmentor

1 code implementation ICCV 2023 Yijiang Li, Xinjiang Wang, Lihe Yang, Litong Feng, Wayne Zhang, Ying Gao

Deep co-training has been introduced to semi-supervised segmentation and achieves impressive results, yet few studies have explored the working mechanism behind it.

Diversity

Multi-metrics adaptively identifies backdoors in Federated learning

1 code implementation ICCV 2023 Siquan Huang, Yijiang Li, Chong Chen, Leyu Shi, Ying Gao

To evaluate the effectiveness of our approach, we conduct comprehensive experiments on different datasets under various attack settings, where our method achieves the best defensive performance.

Federated Learning Privacy Preserving

More than Encoder: Introducing Transformer Decoder to Upsample

no code implementations20 Jun 2021 Yijiang Li, Wentian Cai, Ying Gao, Chengming Li, Xiping Hu

The local and detailed feature from the shallower layer such as boundary and tissue texture is particularly more important in medical segmentation compared with natural image segmentation.

Decoder Image Segmentation +4

Cannot find the paper you are looking for? You can Submit a new open access paper.