Llama-3 8B Instruct 262k is a text generation model developed by the Gradient AI team, extending the context length of Llama-3 8B to over 160K and demonstrating the potential of state-of-the-art large language models in handling long text. This model achieves efficient learning on long texts through proper adjustment of the RoPE theta parameter, combined with NTK-aware interpolation and data-driven optimization techniques. Additionally, it is built upon the EasyContext Blockwise RingAttention library to support scalable and efficient training on high-performance hardware.