/Models/Llama-3.3-70B-Instruct-Turbo
Meta

Llama-3.3-70B-Instruct-Turbo

Lowest Price
$0.10
per 1M tokens
Providers
1
Available
Context
N/A
tokens

Price Comparison

ProviderInput / OutputLatencyStatus
DeepInfraDeepInfraLowest
$0.10/$0.32
...
Verified

About This Model

Llama 3.3-70B Turbo is a highly optimized version of the Llama 3.3-70B model, utilizing FP8 quantization to deliver significantly faster inference speeds with a minor trade-off in accuracy. The model is designed to be helpful, safe, and flexible, with a focus on responsible deployment and mitigating...

Quick Start