Search Results for author: Jialin Li

Found 27 papers, 9 papers with code

AeroGen: Enhancing Remote Sensing Object Detection with Diffusion-Driven Data Generation

1 code implementation23 Nov 2024 Datao Tang, Xiangyong Cao, Xuan Wu, Jialin Li, Jing Yao, Xueru Bai, Deyu Meng

Although existing techniques, e. g., data augmentation and semi-supervised learning, can mitigate this scarcity issue to some extent, they are heavily dependent on high-quality labeled data and perform worse in rare object classes.

Data Augmentation Diversity +2

RoboGSim: A Real2Sim2Real Robotic Gaussian Splatting Simulator

no code implementations18 Nov 2024 Xinhai Li, Jialin Li, Ziheng Zhang, Rui Zhang, Fan Jia, Tiancai Wang, Haoqiang Fan, Kuo-Kun Tseng, Ruiping Wang

To address these limitations, we introduce the RoboGSim, a real2sim2real robotic simulator, powered by 3D Gaussian Splatting and the physics engine.

GAN-Based Architecture for Low-dose Computed Tomography Imaging Denoising

no code implementations14 Nov 2024 Yunuo Wang, Ningning Yang, Jialin Li

Generative Adversarial Networks (GANs) have surfaced as a revolutionary element within the domain of low-dose computed tomography (LDCT) imaging, providing an advanced resolution to the enduring issue of reconciling radiation exposure with image quality.

Denoising SSIM +1

ParaGAN: A Scalable Distributed Training Framework for Generative Adversarial Networks

no code implementations6 Nov 2024 Ziji Shi, Jialin Li, Yang You

Recent advances in Generative Artificial Intelligence have fueled numerous applications, particularly those involving Generative Adversarial Networks (GANs), which are essential for synthesizing realistic photos and videos.

Image Generation

ED-ViT: Splitting Vision Transformer for Distributed Inference on Edge Devices

no code implementations15 Oct 2024 Xiang Liu, Yijun Song, Xia Li, Yifei Sun, Huiying Lan, Zemin Liu, Linshan Jiang, Jialin Li

We conduct extensive experiments on five datasets with three model structures, demonstrating that our approach significantly reduces inference latency on edge devices and achieves a model size reduction of up to 28. 9 times and 34. 1 times, respectively, while maintaining test accuracy comparable to the original Vision Transformer.

MMAD: The First-Ever Comprehensive Benchmark for Multimodal Large Language Models in Industrial Anomaly Detection

1 code implementation12 Oct 2024 Xi Jiang, Jian Li, Hanqiu Deng, Yong liu, Bin-Bin Gao, Yifeng Zhou, Jialin Li, Chengjie Wang, Feng Zheng

In the field of industrial inspection, Multimodal Large Language Models (MLLMs) have a high potential to renew the paradigms in practical applications due to their robust language capabilities and generalization abilities.

Anomaly Detection

CAR: Controllable Autoregressive Modeling for Visual Generation

1 code implementation7 Oct 2024 Ziyu Yao, Jialin Li, Yifeng Zhou, Yong liu, Xi Jiang, Chengjie Wang, Feng Zheng, Yuexian Zou, Lei LI

To the best of our knowledge, we are the first to propose a control framework for pre-trained autoregressive visual generation models.

Safe Time-Varying Optimization based on Gaussian Processes with Spatio-Temporal Kernel

no code implementations26 Sep 2024 Jialin Li, Marta Zagorowska, Giulia De Pasquale, Alisa Rupenyan, John Lygeros

Evaluation on a realistic case study with gas compressors confirms that TVSafeOpt ensures safety when solving time-varying optimization problems with unknown reward and safety functions.

Bayesian Optimization Change Detection +3

Blocks as Probes: Dissecting Categorization Ability of Large Multimodal Models

no code implementations3 Sep 2024 Bin Fu, Qiyang Wan, Jialin Li, Ruiping Wang, Xilin Chen

Categorization, a core cognitive ability in humans that organizes objects based on common features, is essential to cognitive science as well as computer vision.

Question Answering Visual Question Answering

P3P: Pseudo-3D Pre-training for Scaling 3D Masked Autoencoders

no code implementations19 Aug 2024 Xuechao Chen, Ying Chen, Jialin Li, Qiang Nie, Yong liu, QiXing Huang, Yang Li

Inspired by semi-supervised learning leveraging limited labeled data and a large amount of unlabeled data, in this work, we propose a novel self-supervised pre-training framework utilizing the real 3D data and the pseudo-3D data lifted from images by a large depth estimation model.

3D Classification Depth Estimation +1

How Well Do LLMs Identify Cultural Unity in Diversity?

1 code implementation9 Aug 2024 Jialin Li, Junli Wang, Junjie Hu, Ming Jiang

In this study, we introduce a benchmark dataset CUNIT for evaluating decoder-only LLMs in understanding the cultural unity of concepts.

Decoder Diversity +1

Decision Boundary-aware Knowledge Consolidation Generates Better Instance-Incremental Learner

no code implementations5 Jun 2024 Qiang Nie, WeiFu Fu, Yuhuan Lin, Jialin Li, Yifeng Zhou, Yong liu, Lei Zhu, Chengjie Wang

Two issues have to be tackled in the new IIL setting: 1) the notorious catastrophic forgetting because of no access to old data, and 2) broadening the existing decision boundary to new observations because of concept drift.

class-incremental learning Class Incremental Learning +2

M4U: Evaluating Multilingual Understanding and Reasoning for Large Multimodal Models

1 code implementation24 May 2024 Hongyu Wang, Jiayu Xu, Senwei Xie, Ruiping Wang, Jialin Li, Zhaojie Xie, Bin Zhang, Chuyan Xiong, Xilin Chen

In this work, we introduce M4U, a novel and challenging benchmark for assessing the capability of multi-discipline multilingual multimodal understanding and reasoning.

Multimodal Reasoning

Improving Personalisation in Valence and Arousal Prediction using Data Augmentation

no code implementations13 Apr 2024 Munachiso Nwadike, Jialin Li, Hanan Salam

This paper presents our work on an enhanced personalisation strategy, that leverages data augmentation to develop tailored models for continuous valence and arousal prediction.

Data Augmentation Emotion Recognition

LORS: Low-rank Residual Structure for Parameter-Efficient Network Stacking

1 code implementation CVPR 2024 Jialin Li, Qiang Nie, WeiFu Fu, Yuhuan Lin, Guangpin Tao, Yong liu, Chengjie Wang

Deep learning models, particularly those based on transformers, often employ numerous stacked structures, which possess identical architectures and perform similar functions.

Decoder

FedLPA: One-shot Federated Learning with Layer-Wise Posterior Aggregation

1 code implementation30 Sep 2023 Xiang Liu, Liangxi Liu, Feiyang Ye, Yunheng Shen, Xia Li, Linshan Jiang, Jialin Li

Efficiently aggregating trained neural networks from local clients into a global model on a server is a widely researched topic in federated learning.

Federated Learning

Can the Query-based Object Detector Be Designed with Fewer Stages?

no code implementations28 Sep 2023 Jialin Li, WeiFu Fu, Yuhuan Lin, Qiang Nie, Yong liu

Query-based object detectors have made significant advancements since the publication of DETR.

Decoder

A Survey on Personalized Affective Computing in Human-Machine Interaction

no code implementations1 Apr 2023 Jialin Li, Alia Waleed, Hanan Salam

In this paper, we discuss the need for personalization in affective and personality computing (hereinafter referred to as affective computing).

Survey

TAP: Accelerating Large-Scale DNN Training Through Tensor Automatic Parallelisation

no code implementations1 Feb 2023 Ziji Shi, Le Jiang, Ang Wang, Jie Zhang, Xianyan Jia, Yong Li, Chencan Wu, Jialin Li, Wei Lin

However, finding a suitable model parallel schedule for an arbitrary neural network is a non-trivial task due to the exploding search space.

MapQA: A Dataset for Question Answering on Choropleth Maps

1 code implementation15 Nov 2022 Shuaichen Chang, David Palzer, Jialin Li, Eric Fosler-Lussier, Ningchuan Xiao

Our experimental results show that V-MODEQA has better overall performance and robustness on MapQA than the state-of-the-art ChartQA and VQA algorithms by capturing the unique properties in map question answering.

Question Answering Visual Question Answering

Harmonia: Near-Linear Scalability for Replicated Storage with In-Network Conflict Detection

no code implementations18 Apr 2019 Hang Zhu, Zhihao Bai, Jialin Li, Ellis Michael, Dan Ports, Ion Stoica, Xin Jin

Experimental results show that Harmonia improves the throughput of these protocols by up to 10X for a replication factor of 10, providing near-linear scalability up to the limit of our testbed.

Distributed, Parallel, and Cluster Computing

Guaranteed Simultaneous Asymmetric Tensor Decomposition via Orthogonalized Alternating Least Squares

no code implementations25 May 2018 Furong Huang, Jialin Li, Xuchen You

We propose a Slicing Initialized Alternating Subspace Iteration (s-ASI) method that is guaranteed to recover top $r$ components ($\epsilon$-close) simultaneously for (a)symmetric tensors almost surely under the noiseless case (with high probability for a bounded noise) using $O(\log(\log \frac{1}{\epsilon}))$ steps of tensor subspace iterations.

Tensor Decomposition

Cannot find the paper you are looking for? You can Submit a new open access paper.