Capabilities
updated
Training-Free Long-Context Scaling of Large Language Models
Paper
• 2402.17463
• Published
• 24
Evaluating Very Long-Term Conversational Memory of LLM Agents
Paper
• 2402.17753
• Published
• 19
Resonance RoPE: Improving Context Length Generalization of Large
Language Models
Paper
• 2403.00071
• Published
• 24
BurstAttention: An Efficient Distributed Attention Framework for
Extremely Long Sequences
Paper
• 2403.09347
• Published
• 22
Data Engineering for Scaling Language Models to 128K Context
Paper
• 2402.10171
• Published
• 25
Long-context LLMs Struggle with Long In-context Learning
Paper
• 2404.02060
• Published
• 37
Leave No Context Behind: Efficient Infinite Context Transformers with
Infini-attention
Paper
• 2404.07143
• Published
• 111
LLoCO: Learning Long Contexts Offline
Paper
• 2404.07979
• Published
• 22
Megalodon: Efficient LLM Pretraining and Inference with Unlimited
Context Length
Paper
• 2404.08801
• Published
• 66
Infinite-LLM: Efficient LLM Service for Long Context with DistAttention
and Distributed KVCache
Paper
• 2401.02669
• Published
• 17
Make Your LLM Fully Utilize the Context
Paper
• 2404.16811
• Published
• 55
Summary of a Haystack: A Challenge to Long-Context LLMs and RAG Systems
Paper
• 2407.01370
• Published
• 89
s1: Simple test-time scaling
Paper
• 2501.19393
• Published
• 124