Models/Deepseek V3
DeepSeekDeepSeek / Deepseek V3
Released: 12/26/2024
texttext
Input: $0.75 / Output: $3.00

Deepseek V3 is an LLM that employs a Mixture-of-Experts (MoE) architecture. It excels in handling extremely long input sequences—up to 128,000 tokens—while delivering efficient inference and strong performance across diverse domains, including complex reasoning, mathematics, and code generation.

Some other noteworthy features of Deepseek V3 include its comprehensive knowledge due to pre-training on 14.8 trillion high-quality tokens and enhanced inference speed, partly enabled by advanced Multi-Token Prediction techniques.

MetricValue
Parameter Count671 billion
Mixture of ExpertsYes
Active Parameter Count37 billion
Context Length128,000 tokens
MultilingualYes
Quantized*No

*Quantization is specific to the inference provider and the model may be offered with different quantization levels by other providers.

DeepSeek models available on Oxen.ai
ModalityPrice (1M tokens)
ModelInference providerInputOutputInputOutput
Fireworks AIFireworks AI
texttext$3.00$8.00
DeepSeekDeepSeek
texttext$0.55$2.19
Together.aiTogether.ai
texttext$7.00$7.00
GroqGroq
texttext$0.59$0.79
Fireworks AIFireworks AI
texttext$0.75$3.00
DeepSeekDeepSeek
texttext$0.27$1.10
Together.aiTogether.ai
texttext$1.25$1.25
See all models available on Oxen.ai