LightRAG: Simple and Fast Retrieval-Augmented Generation

hkuds/lightrag 8 Oct 2024

Retrieval-Augmented Generation (RAG) systems enhance large language models (LLMs) by integrating external knowledge sources, enabling more accurate and contextually relevant responses tailored to user needs.

Information Retrieval RAG +1

8,103
0.62 stars / hour

In-Context LoRA for Diffusion Transformers

ali-vilab/In-Context-LoRA 31 Oct 2024

While task-specific in terms of tuning data, our framework remains task-agnostic in architecture and pipeline, offering a powerful tool for the community and providing valuable insights for further research on product-level task-agnostic generation systems.

Image Generation

508
0.61 stars / hour

TableGPT2: A Large Multimodal Model with Tabular Data Integration

tablegpt/tablegpt-agent 4 Nov 2024

In response, we introduce TableGPT2, a model rigorously pre-trained and fine-tuned with over 593. 8K tables and 2. 36M high-quality query-table-output tuples, a scale of table-related data unprecedented in prior research.

Benchmarking Data Integration

197
0.61 stars / hour

Extended Agriculture-Vision: An Extension of a Large Aerial Image Dataset for Agricultural Pattern Analysis

jingwu6/extended-agriculture-vision-dataset 4 Mar 2023

First, we generate and release an improved version of the Agriculture-Vision dataset (Chiu et al., 2020b) to include raw, full-field imagery for greater experimental flexibility.

Benchmarking Contrastive Learning +2

203
0.56 stars / hour

QServe: W4A8KV4 Quantization and System Co-design for Efficient LLM Serving

mit-han-lab/deepcompressor 7 May 2024

The key insight driving QServe is that the efficiency of LLM serving on GPUs is critically influenced by operations on low-throughput CUDA cores.

Language Modelling Large Language Model +1

193
0.55 stars / hour

TokenFormer: Rethinking Transformer Scaling with Tokenized Model Parameters

haiyang-w/tokenformer 30 Oct 2024

By treating model parameters as tokens, we replace all the linear projections in Transformers with our token-parameter attention layer, where input tokens act as queries and model parameters as keys and values.

238
0.52 stars / hour

LLM$\times$MapReduce: Simplified Long-Sequence Processing using Large Language Models

thunlp/llmxmapreduce 12 Oct 2024

The proposed LLM$\times$MapReduce framework splits the entire document into several chunks for LLMs to read and then aggregates the intermediate answers to produce the final output.

document understanding

105
0.51 stars / hour

OpenHands: An Open Platform for AI Software Developers as Generalist Agents

opendevin/opendevin 23 Jul 2024

OpenDevin), a platform for the development of powerful and flexible AI agents that interact with the world in similar ways to those of a human developer: by writing code, interacting with a command line, and browsing the web.

36,152
0.50 stars / hour

Hunyuan-Large: An Open-Source MoE Model with 52 Billion Activated Parameters by Tencent

tencent/hunyuan-large 4 Nov 2024

In this paper, we introduce Hunyuan-Large, which is currently the largest open-source Transformer-based mixture of experts model, with a total of 389 billion parameters and 52 billion activation parameters, capable of handling up to 256K tokens.

Logical Reasoning Mathematical Problem-Solving

1,029
0.45 stars / hour

Training-free Regional Prompting for Diffusion Transformers

instantX-research/Regional-Prompting-FLUX 4 Nov 2024

Diffusion models have demonstrated excellent capabilities in text-to-image generation.

Text-to-Image Generation

315
0.44 stars / hour