deepseek-chat

by deepseek

Pricing

Input $0.11 / 1M tokens
Output $0.44 / 1M tokens

DeepSeek Chat (DeepSeek v3) Release by DeepSeek


Overview

DeepSeek Chat (DeepSeek v3) is a large-scale, mixture-of-experts AI model released in March 2025, featuring a massive 128K token context window and strong capabilities in reasoning, coding, and multilingual tasks.

  • Release Date: March 24, 2025
  • Architecture: Mixture-of-Experts (MoE), 671B parameters total, 37B active experts per token
  • Context Length: 128,000 tokens
  • Maximum Output Length: 8,000 tokens
  • Inference Speed: ~60 tokens per second
  • Latency: ~3.03 seconds to first token
  • Training Data Cutoff: July 1, 2024

Features

  • Enhanced logical reasoning and coding capabilities
  • Multilingual support
  • Advanced integration with external tools and APIs
  • High throughput with MoE architecture

Benchmarks

Task DeepSeek Chat Claude 3 Opus Gemini 2.5 Flash
MMLU (Reasoning) 79.2% 86.8% 77.9%
MGSM (Math) 81.3% 60.1% 79.7%
HumanEval (Coding) 76.5% 84.9% 74.1%
MMMU (Multimodal) 57.3% 54.9% 56.8%

Pricing

Type Price per 1M tokens
Input $0.27
Output $1.10

Use Cases

  • Real-time chatbots and customer support
  • Educational tutoring and study assistance
  • Content generation and summarization
  • Software development and debugging
  • Data analysis and insights

Safety and Stability

  • Open source under MIT license
  • No live internet connection (offline model)
  • Strong instruction following and robustness
  • Training data cutoff: July 1, 2024

Limitations

  • Only text input/output supported (no multimodal yet)
  • Possible numeric precision inconsistencies
  • Latency higher compared to smaller models due to size

License

DeepSeek Chat (DeepSeek v3) is licensed under the MIT License and allows commercial use.