Search Results for author: Haowen Hou

Found 12 papers, 6 papers with code

Delta-WKV: A Novel Meta-in-Context Learner for MRI Super-Resolution

no code implementations28 Feb 2025 Rongchang Lu, Bingcheng Liao, Haowen Hou, Jiahang Lv, Xin Hai

Magnetic Resonance Imaging (MRI) Super-Resolution (SR) addresses the challenges such as long scan times and expensive equipment by enhancing image resolution from low-quality inputs acquired in shorter scan times in clinical settings.

In-Context Learning SSIM +1

RWKV-UI: UI Understanding with Enhanced Perception and Reasoning

no code implementations6 Feb 2025 Jiaxi Yang, Haowen Hou

Additionally, we design a visual prompt based on the Chain-of-Thought(CoT) mechanism, which enhances the model's ability to understand and reason about webpage content through reasoning chains.

Language Modeling Language Modelling

ImDy: Human Inverse Dynamics from Imitated Observations

no code implementations23 Oct 2024 Xinpeng Liu, Junxuan Liang, Zili Lin, Haowen Hou, Yong-Lu Li, Cewu Lu

In light of this, we devise an efficient data collection pipeline with state-of-the-art motion imitation algorithms and physics simulators, resulting in a large-scale human inverse dynamics benchmark as Imitated Dynamics (ImDy).

VisualRWKV-HD and UHD: Advancing High-Resolution Processing for Visual Language Models

no code implementations15 Oct 2024 Zihang Li, Haowen Hou

Accurately understanding complex visual information is crucial for visual language models (VLMs).

Enhancing and Accelerating Large Language Models via Instruction-Aware Contextual Compression

1 code implementation28 Aug 2024 Haowen Hou, Fei Ma, Binwen Bai, Xinxin Zhu, Fei Yu

Large Language Models (LLMs) have garnered widespread attention due to their remarkable performance across various tasks.

VisualRWKV: Exploring Recurrent Neural Networks for Visual Language Models

1 code implementation19 Jun 2024 Haowen Hou, Peigen Zeng, Fei Ma, Fei Richard Yu

Visual Language Models (VLMs) have rapidly progressed with the recent success of large language models.

Language Modeling Language Modelling

RWKV-TS: Beyond Traditional Recurrent Neural Network for Time Series Tasks

2 code implementations17 Jan 2024 Haowen Hou, F. Richard Yu

The combination of competitive performance, low latency, and efficient memory usage positions RWKV-TS as a promising avenue for future research in time series tasks.

Computational Efficiency Time Series +1

Revisit Human-Scene Interaction via Space Occupancy

no code implementations5 Dec 2023 Xinpeng Liu, Haowen Hou, Yanchao Yang, Yong-Lu Li, Cewu Lu

High-quality data with simultaneously captured human and 3D environments is hard to acquire, resulting in limited data diversity and complexity.

BagFormer: Better Cross-Modal Retrieval via bag-wise interaction

no code implementations29 Dec 2022 Haowen Hou, Xiaopeng Yan, Yigeng Zhang, Fengzong Lian, Zhanhui Kang

In the field of cross-modal retrieval, single encoder models tend to perform better than dual encoder models, but they suffer from high latency and low throughput.

Cross-Modal Retrieval Retrieval

Sams-Net: A Sliced Attention-based Neural Network for Music Source Separation

1 code implementation12 Sep 2019 Tingle Li, Jia-Wei Chen, Haowen Hou, Ming Li

Convolutional Neural Network (CNN) or Long short-term memory (LSTM) based models with the input of spectrogram or waveforms are commonly used for deep learning based audio source separation.

Audio Source Separation Music Source Separation

Cannot find the paper you are looking for? You can Submit a new open access paper.