is our sole official account on Twitter/X.
Any accounts:
- representing us
- using identical avatars
- using similar names
are impersonations.
Please stay vigilant to avoid being misled!
DeepSeek-R1 is here!
Performance on par with OpenAI-o1
Fully open-source model & technical report
MIT licensed: Distill & commercialize freely!
Website & API are live now! Try DeepThink at http://chat.deepseek.com today!
1/n
License Update!
DeepSeek-R1 is now MIT licensed for clear open access
Open for the community to leverage model weights & outputs
API outputs can now be used for fine-tuning & distillation
3/n
DeepSeek-R1-Lite-Preview is now live: unleashing supercharged reasoning power!
o1-preview-level performance on AIME & MATH benchmarks.
Transparent thought process in real-time.
Open-source models & API coming soon!
Try it now at http://chat.deepseek.com#DeepSeek
Bonus: Open-Source Distilled Models!
Distilled from DeepSeek-R1, 6 small models fully open-sourced
32B & 70B models on par with OpenAI-o1-mini
Empowering the open-source community
Pushing the boundaries of **open AI**!
2/n
Introducing DeepSeek App!
Powered by world-class DeepSeek-V3
FREE to use with seamless interaction
Now officially available on App Store & Google Play & Major Android markets
Download now: https://download.deepseek.com/app/ 1/3
DeepSeek-Coder-V2: First Open Source Model Beats GPT4-Turbo in Coding and Math
> Excels in coding and math, beating GPT4-Turbo, Claude3-Opus, Gemini-1.5Pro, Codestral.
> Supports 338 programming languages and 128K context length.
> Fully open-sourced with two sizes: 230B (also
DeepSeek-V2.5-1210: The Grand Finale Internet Search is now live on the web! Visit https://chat.deepseek.com and toggle “Internet Search” for real-time answers. (1/3)
API Access & Pricing
Use DeepSeek-R1 by setting model=deepseek-reasoner
$0.14 / million input tokens (cache hit)
$0.55 / million input tokens (cache miss)
$2.19 / million output tokens
API guide: https://api-docs.deepseek.com/guides/reasoning_model… 5/n
DeepSeek-VL2 is here! Our next-gen vision-language model enters the MoE era.
DeepSeek-MoE arch + dynamic image tilling
3B/16B/27B sizes for flexible use
Outstanding performance across all benchmarks
1/n
API Pricing Update
Until Feb 8: same as V2!
From Feb 8 onwards:
Input: $0.27/million tokens ($0.07/million tokens with cache hits)
Output: $1.10/million tokens
Still the best value in the market!
3/n
Launching DeepSeek-V2: The Cutting-Edge Open-Source MoE Model!
Highlights:
> Places top 3 in AlignBench, surpassing GPT-4 and close to GPT-4-Turbo.
> Ranks top-tier in MT-Bench, rivaling LLaMA3-70B and outperforming Mixtral 8x22B.
> Specializes in math, code and reasoning.
DeepSeekMath: Approaching Mathematical Reasoning Capability of GPT-4 with a 7B Model.
Highlights:
- Continue pre-training DeepSeek-Coder-Base-v1.5 7B with 120B math tokens from Common Crawl.
- Introduce GRPO, a variant of PPO, that enhances mathematical reasoning and reduces
Open-source spirit + Longtermism to inclusive AGI
DeepSeek’s mission is unwavering. We’re thrilled to share our progress with the community and see the gap between open and closed models narrowing.
This is just the beginning! Look forward to multimodal support and
DeepSeek has not issued any cryptocurrency. Currently, there is only one official account on the Twitter platform. We will not contact anyone through other accounts.Please stay vigilant and guard against potential scams.
Introducing Janus: a revolutionary autoregressive framework for multimodal AI!
By decoupling visual encoding & unifying them with a single transformer, it outperforms previous models in both understanding & generation.
Powerful, simple, flexible, & next-gen ready!
Respect to Artifacts of Claude 3.5 Sonnet!
DeepSeek-Coder-V2 can do the same cool stuff directly in your browser.
Visit http://coder.deepseek.com -> select "Coder V2" -> input prompt -> click “Run HTML” to see the magic happen!
#DeepSeekCoder#Claude
Exciting news! We’ve officially launched DeepSeek-V2.5 – a powerful combination of DeepSeek-V2-0628 and DeepSeek-Coder-V2-0724! Now, with enhanced writing, instruction-following, and human preference alignment, it’s available on Web and API. Enjoy seamless Function Calling,
Exciting news! DeepSeek API now launches context caching on disk, with no code changes required! This new feature automatically caches frequently referenced contexts on distributed storage, slashing API costs by up to 90%. For a 128K prompt with high reference, the first token
After 3 months, the AI Mathematical Olympiad (AIMO) on Kaggle has announced the winners!
We're thrilled to see the Top 4 teams all chose DeepSeekMath-7B as their base model, with Numina