Search Results for author: Xingtong Ge

Found 5 papers, 2 papers with code

Task-Aware Encoder Control for Deep Video Compression

no code implementations7 Apr 2024 Xingtong Ge, Jixiang Luo, Xinjie Zhang, Tongda Xu, Guo Lu, Dailan He, Jing Geng, Yan Wang, Jun Zhang, Hongwei Qin

Prior research on deep video compression (DVC) for machine tasks typically necessitates training a unique codec for each specific task, mandating a dedicated decoder per task.

Video Compression

GaussianImage: 1000 FPS Image Representation and Compression by 2D Gaussian Splatting

1 code implementation13 Mar 2024 Xinjie Zhang, Xingtong Ge, Tongda Xu, Dailan He, Yan Wang, Hongwei Qin, Guo Lu, Jing Geng, Jun Zhang

In response, we propose a groundbreaking paradigm of image representation and compression by 2D Gaussian Splatting, named GaussianImage.

Quantization

Content-aware Masked Image Modeling Transformer for Stereo Image Compression

no code implementations13 Mar 2024 Xinjie Zhang, Shenyuan Gao, Zhening Liu, Jiawei Shao, Xingtong Ge, Dailan He, Tongda Xu, Yan Wang, Jun Zhang

Existing learning-based stereo image codec adopt sophisticated transformation with simple entropy models derived from single image codecs to encode latent representations.

Image Compression

Boosting Neural Representations for Videos with a Conditional Decoder

1 code implementation28 Feb 2024 Xinjie Zhang, Ren Yang, Dailan He, Xingtong Ge, Tongda Xu, Yan Wang, Hongwei Qin, Jun Zhang

Implicit neural representations (INRs) have emerged as a promising approach for video storage and processing, showing remarkable versatility across various video tasks.

Preprocessing Enhanced Image Compression for Machine Vision

no code implementations12 Jun 2022 Guo Lu, Xingtong Ge, Tianxiong Zhong, Jing Geng, Qiang Hu

Specifically, we propose a neural preprocessing module before the encoder to maintain the useful semantic information for the downstream tasks and suppress the irrelevant information for bitrate saving.

Image Compression object-detection +2

Cannot find the paper you are looking for? You can Submit a new open access paper.