Long-range modeling

45 papers with code • 2 benchmarks • 4 datasets

A new task for testing the long-sequence modeling capabilities and efficiency of language models.

Image credit: SCROLLS: Standardized CompaRison Over Long Language Sequences

Libraries

Use these libraries to find Long-range modeling models and implementations
2 papers
56

Most implemented papers

T-former: An Efficient Transformer for Image Inpainting

dengyecode/t-former_image_inpainting 12 May 2023

And based on this attention, a network called $T$-former is designed for image inpainting.

MambaMorph: a Mamba-based Framework for Medical MR-CT Deformable Registration

guo-stone/mambamorph 25 Jan 2024

Capturing voxel-wise spatial correspondence across distinct modalities is crucial for medical image analysis.

VM-UNet: Vision Mamba UNet for Medical Image Segmentation

jcruan519/vm-unet 4 Feb 2024

To our best knowledge, this is the first medical image segmentation model constructed based on the pure SSM-based model.

V4D:4D Convolutional Neural Networks for Video-level Representation Learning

MalongTech/research-v4d 18 Feb 2020

Most existing 3D CNNs for video representation learning are clip-based methods, and thus do not consider video-level temporal evolution of spatio-temporal features.

DSANet: Dynamic Segment Aggregation Network for Video-Level Representation Learning

whwu95/DSANet 25 May 2021

Long-range and short-range temporal modeling are two complementary and crucial aspects of video recognition.

Image Super-Resolution With Non-Local Sparse Attention

HarukiYqM/Non-Local-Sparse-Attention CVPR 2021

NLSA is designed to retain long-range modeling capability from NL operation while enjoying robustness and high-efficiency of sparse representation.

Sparse Factorization of Large Square Matrices

ruslankhalitov/sparsefactorization 16 Sep 2021

The sparse factorization method is tested for a variety of synthetic and real-world square matrices.

LongT5: Efficient Text-To-Text Transformer for Long Sequences

google-research/longt5 Findings (NAACL) 2022

Recent work has shown that either (1) increasing the input length or (2) increasing model size can improve the performance of Transformer-based neural models.

Classification of Long Sequential Data using Circular Dilated Convolutional Neural Networks

leicheng-no/cdil-cnn 6 Jan 2022

Recurrent Neural Networks, Transformers, and Convolutional Neural Networks are three major techniques for learning from sequential data.