Sequence Parallel Attention for Long Context LLM Model Training and Inference
-
Updated
May 16, 2024 - Python
Sequence Parallel Attention for Long Context LLM Model Training and Inference
Packaged Ring Attention with Blockwise Transformers for Near-Infinite Context implemented in Jax + Flax.
Add a description, image, and links to the ring-attention topic page so that developers can more easily learn about it.
To associate your repository with the ring-attention topic, visit your repo's landing page and select "manage topics."