Back to Chatgpt On Wechat

DeepSeek

docs/en/models/deepseek.mdx

2.0.82.0 KB
Original Source

Option 1: Native integration (recommended):

json
{
  "model": "deepseek-v4-flash",
  "deepseek_api_key": "YOUR_API_KEY"
}
ParameterDescription
modelSupports deepseek-v4-flash (default) and deepseek-v4-pro
deepseek_api_keyCreate at DeepSeek Platform
deepseek_api_baseOptional, defaults to https://api.deepseek.com/v1. Can be changed to a third-party proxy

Model Selection

ModelUse Case
deepseek-v4-flashDefault: fast and cost-effective
deepseek-v4-proStronger on complex tasks

Thinking Mode

The V4 series (deepseek-v4-flash / deepseek-v4-pro) supports an explicit "thinking mode": the model emits a chain-of-thought (reasoning_content) before the final answer to improve answer quality.

Toggle

Controlled by the global enable_thinking setting:

json
{
  "enable_thinking": true
}
  • true: thinking is on across all channels. The Web console renders the reasoning trace; IM channels (WeChat / WeCom / DingTalk / Feishu) don't render it but still benefit from higher answer quality.
  • false: thinking off, faster responses with lower first-token latency.

Notes

  • Sampling parameters: under thinking mode, temperature, top_p, presence_penalty, and frequency_penalty are silently ignored by the server (no error). CowAgent skips sending them automatically.
  • Multi-turn tool calls: once the history contains any tool-call turn, DeepSeek requires reasoning_content on every assistant message. CowAgent handles the round-trip automatically, including across mid-session toggles of the thinking switch.
<Tip> Start with `deepseek-v4-flash`; switch to `deepseek-v4-pro` for harder tasks; enable `enable_thinking` when you want deeper reasoning. </Tip>

Option 2: OpenAI-compatible configuration:

json
{
  "model": "deepseek-v4-flash",
  "bot_type": "openai",
  "open_ai_api_key": "YOUR_API_KEY",
  "open_ai_api_base": "https://api.deepseek.com/v1"
}