DeepSeek_R1使用指南

DeepSeek案例介绍
大家好,欢迎来到AI使用技巧课堂!
相信大家这两天已经被deepseek炸屏了。
deepseek
搜索
“DeepSeek接班OpenAl",R1推理模型让Al圈爆了

大家好,欢迎来到AI使用技巧课堂!
相信大家这两天已经被deepseek炸屏了。
“DeepSeek接班OpenAl",R1推理模型让Al圈爆了

大家好,欢迎来到AI使用技巧课堂!
相信大家这两天已经被deepseek炸屏了。
DeepSeek接班OpenAl,R1推理模型让Al圈爆了
注册一个 deepseek 账号并获取密钥(api key),后面 anythingLLM 会使用这个密钥接入 deepseek

DeepSeek-AI
We present DeepSeek-V3, a strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token. To achieve efficient inference and cost-effective training, DeepSeek-V3 adopts Multi-head Latent Attention (MLA) and DeepSeekMoE architectures, which were thoroughly validated in DeepSeek-V2. Furthermore, DeepSeek-V3 pioneers an auxiliary-loss-free strategy for load balancing and sets a multi-token prediction training objective for stronger performance. We pre-train DeepSeek-V3 on 14.8 trillion diverse and high-quality tokens, followed by Supervised Fine-Tuning and Reinforcement Learning stages to fully harness its capabilities. Comprehensive evaluations reveal that DeepSeek-V3 outperforms other open-source models and achieves performance comparable to leading closed-source models. Despite its excellent performance, DeepSeek-V3 requires only 2.788M H800 GPU hours for its full training. In addition, its training process is remarkably stable. Throughout the entire training process, we did not experience any irrecoverable loss spikes or perform any rollbacks. The model checkpoints are available at https://github.com/deepseek-ai/DeepSeek-V3.
DeepSeek-AI
We present DeepSeek-V3, a strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token. To achieve efficient inference and cost-effective training, DeepSeek-V3 adopts Multi-head Latent Attention (MLA) and DeepSeekMoE architectures, which were thoroughly validated in DeepSeek-V2. Furthermore, DeepSeek-V3 pioneers an auxiliary-loss-free strategy for load balancing and sets a multi-token prediction training objective for stronger performance. We pre-train DeepSeek-V3 on 14.8 trillion diverse and high-quality tokens, followed by Supervised Fine-Tuning and Reinforcement Learning stages to fully harness its capabilities. Comprehensive evaluations reveal that DeepSeek-V3 outperforms other open-source models and achieves performance comparable to leading closed-source models. Despite its excellent performance, DeepSeek-V3 requires only 2.788M H800 GPU hours for its full training. In addition, its training process is remarkably stable. Throughout the entire training process, we did not experience any irrecoverable loss spikes or perform any rollbacks. The model checkpoints are available at https://github.com/deepseek-ai/DeepSeek-V3.
2025 Deepseek 爆火全球,会用的人说巨好用,差距就在提问方式,教你用 1个公式让 Deepseek 变身“职场军师”“学霸外挂”“流量引擎”。
凝炼出一套方法论,按万能公式「4 步提问法」
DeepSeek搞钱教程
(0基础入门)

微信扫码,长期有效
为了让我的回答更准确和有用,你可以通过以下方式优化提问。我通过几个例子为你说明不同提问方式的区别:
$\star\star\mathbb{1}$ . 明确问题类型 $^+$ 具体细节**
好例子:
1.访问官网:在浏览器输入「www.deepseek.com」 (就像打开微信一样简单)
2.注册账号:点击右上角「笑脸图标」 $\rightarrow$ 选择 「邮箱/手机注册」 (建议使用常用邮箱)
(适合零基础小白)

微信扫码,长期有效