Llama 3.1 8B Instruct Turbo
meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo
Llama 3.1 8B model utilizes FP8 quantization, supporting up to 131,072 context tokens, making it a standout in open-source models, excelling in complex tasks and outperforming many industry benchmarks.
128K