Fine-tuning Llama 3.1 with Long Sequences
databricks
SEPTEMBER 19, 2024
Mosaic AI Model Training now supports fine-tuning up to 131K context length for Llama 3.1 models. More efficient training at long sequence lengths is made possible by several optimizations highlighted in this post.
Let's personalize your content