paper-conference

Effective Long-Context Scaling of Foundation Models
We present a series of long-context LLMs that support effective context windows of up to 32,768 tokens. Our model series are built …
Effective Long-Context Scaling of Foundation Models