no code implementations • COLING 2022 • Yen-Hao Huang, Yi-Hsin Chen, Yi-Shin Chen
In this work, we propose a simple yet effective unified model, coined ConTextING, with a joint training mechanism to learn from both document embeddings and contextual word interactions simultaneously.
no code implementations • 29 Jul 2024 • Chia-Hao Kao, Cheng Chien, Yu-Jen Tseng, Yi-Hsin Chen, Alessandro Gnutti, Shao-Yuan Lo, Wen-Hsiao Peng, Riccardo Leonardi
MLLMs have extended the success of large language models to modalities (e. g. images) beyond text, but their billion scale hinders deployment on resource-constrained end devices.
no code implementations • 20 Feb 2024 • Yi-Hsin Chen, Kuan-Wei Ho, Shiau-Rung Tsai, Guan-Hsun Lin, Alessandro Gnutti, Wen-Hsiao Peng, Riccardo Leonardi
Instead of training separate decoders for these tasks, we incorporate two add-on modules to adapt a pre-trained image decoder from performing the standard image reconstruction to joint decoding and denoising.
no code implementations • 12 Jan 2024 • Alessandro Gnutti, Stefano Della Fiore, Mattia Savardi, Yi-Hsin Chen, Riccardo Leonardi, Wen-Hsiao Peng
In this paper, we introduce a novel direction that harnesses LiDAR depth maps to enhance the compression of the corresponding RGB camera images.
no code implementations • 25 Dec 2023 • Yi-Hsin Chen, Hong-Sheng Xie, Cheng-Wei Chen, Zong-Lin Gao, Martin Benjak, Wen-Hsiao Peng, Jörn Ostermann
Conditional coding has lately emerged as the mainstream approach to learned video compression.
no code implementations • 22 Sep 2023 • Chia-Hao Kao, Yi-Hsin Chen, Cheng Chien, Wei-Chen Chiu, Wen-Hsiao Peng
This paper presents a Transformer-based image compression system that allows for a variable image quality objective according to the user's preference.
1 code implementation • ICCV 2023 • Si-Cun Chen, Yi-Hsin Chen, Yen-Yu Lin, Wen-Hsiao Peng
We motivate the use of forward motion from the perspective of learning individual motion trajectories, as opposed to learning a mixture of motion trajectories with backward motion.
1 code implementation • ICCV 2023 • Yi-Hsin Chen, Ying-Chieh Weng, Chia-Hao Kao, Cheng Chien, Wei-Chen Chiu, Wen-Hsiao Peng
This work aims for transferring a Transformer-based image compression codec from human perception to machine perception without fine-tuning the codec.
1 code implementation • 18 May 2023 • Chia-Hao Kao, Ying-Chieh Weng, Yi-Hsin Chen, Wei-Chen Chiu, Wen-Hsiao Peng
Our prompt generation networks generate content-adaptive tokens according to the input image, an ROI mask, and a rate parameter.
no code implementations • 13 Feb 2023 • Chih-Hsuan Lin, Yi-Hsin Chen, Wen-Hsiao Peng
This paper introduces an online motion rate adaptation scheme for learned video compression, with the aim of achieving content-adaptive coding on individual test sequences to mitigate the domain gap between training and test data.
1 code implementation • 5 Sep 2022 • Mu-Jung Chen, Yi-Hsin Chen, Wen-Hsiao Peng
Our B*-frames allow greater flexibility in specifying the group-of-pictures (GOP) structure by reusing the B-frame codec to mimic P-frame coding, without the need for an additional, separate P-frame codec.
1 code implementation • CVPR 2021 • Yan-Cheng Huang, Yi-Hsin Chen, Cheng-You Lu, Hui-Po Wang, Wen-Hsiao Peng, Ching-Chun Huang
Our Long Short-Term Memory Video Rescaling Network (LSTM-VRN) leverages temporal information in the low-resolution video to form an explicit prediction of the missing high-frequency information for upscaling.
1 code implementation • 15 Dec 2020 • Cheng-Hsun Lei, Yi-Hsin Chen, Wen-Hsiao Peng, Wei-Chen Chiu
In this paper, we address the problem of distillation-based class-incremental learning with a single head.
no code implementations • 9 Dec 2020 • Chia-Yu Hsu, Yu-Sheng Wang, Jia-Mou Chen, Fu-Chen Huang, Yi-Ting Ke, Emily Kay Huang, Weilun Hung, Kai-Lin Chao, Shih-Si Hsiao, Yi-Hsin Chen, Chih-Sung Chuu, Ying-Cheng Chen, Yong-Fan Chen, Ite A. Yu
The generation rate per linewidth of the 610-kHz biphoton source is 1, 500 pairs/(s$\cdot$MHz), which is the best result of all the sub-MHz biphoton sources in the literature.
Quantum Physics
1 code implementation • 17 Aug 2019 • Yen-Hao Huang, Ssu-Rui Lee, Mau-Yun Ma, Yi-Hsin Chen, Ya-Wen Yu, Yi-Shin Chen
By the nature of the framework of BERT, a two-sentence structure, we adapt BERT to continues dialogue emotion prediction tasks, which rely heavily on the sentence-level context-aware understanding.
no code implementations • 17 Jul 2019 • Yen-Hao Huang, Yi-Hsin Chen, Fernando Henrique Calderon Alvarado, Ssu-Rui Lee, Shu-I Wu, Yuwen Lai, Yi-Shin Chen
The factors examined are gender differences, syntactic patterns, and bipolar recognition performance.
9 code implementations • ICCV 2017 • Yi-Hsin Chen, Wei-Yu Chen, Yu-Ting Chen, Bo-Cheng Tsai, Yu-Chiang Frank Wang, Min Sun
Despite the recent success of deep-learning based semantic segmentation, deploying a pre-trained road scene segmenter to a city whose images are not presented in the training set would not achieve satisfactory performance due to dataset biases.