DeepSeek / Deepseek V3
Released: 12/26/2024texttext
Input: $0.75 / Output: $3.00
Deepseek V3 is an LLM that employs a Mixture-of-Experts (MoE) architecture. It excels in handling extremely long input sequences—up to 128,000 tokens—while delivering efficient inference and strong performance across diverse domains, including complex reasoning, mathematics, and code generation.
Some other noteworthy features of Deepseek V3 include its comprehensive knowledge due to pre-training on 14.8 trillion high-quality tokens and enhanced inference speed, partly enabled by advanced Multi-Token Prediction techniques.
Metric | Value |
---|---|
Parameter Count | 671 billion |
Mixture of Experts | Yes |
Active Parameter Count | 37 billion |
Context Length | 128,000 tokens |
Multilingual | Yes |
Quantized* | No |
*Quantization is specific to the inference provider and the model may be offered with different quantization levels by other providers.
DeepSeek models available on Oxen.ai
Modality | Price (1M tokens) | ||||
---|---|---|---|---|---|
Model | Inference provider | Input | Output | Input | Output |
![]() | text | text | $3.00 | $8.00 | |
text | text | $0.55 | $2.19 | ||
text | text | $7.00 | $7.00 | ||
![]() | text | text | $0.59 | $0.79 | |
![]() | text | text | $0.75 | $3.00 | |
text | text | $0.27 | $1.10 | ||
text | text | $1.25 | $1.25 |