A curated list of resources about long-context in large-language models and video understanding.
- Recurrent Memory Transformer
- LongNet: Scaling Transformers to 1,000,000,000 Tokens
- Focused transformer
- Augmenting Language Models with Long-Term Memory
- Memorizing Transformers
This is a work in progress. Contributions welcome!