docs/en/models/deepseek.mdx
Option 1: Native integration (recommended):
{
"model": "deepseek-v4-flash",
"deepseek_api_key": "YOUR_API_KEY"
}
| Parameter | Description |
|---|---|
model | Supports deepseek-v4-flash (default) and deepseek-v4-pro |
deepseek_api_key | Create at DeepSeek Platform |
deepseek_api_base | Optional, defaults to https://api.deepseek.com/v1. Can be changed to a third-party proxy |
| Model | Use Case |
|---|---|
deepseek-v4-flash | Default: fast and cost-effective |
deepseek-v4-pro | Stronger on complex tasks |
The V4 series (deepseek-v4-flash / deepseek-v4-pro) supports an explicit "thinking mode": the model emits a chain-of-thought (reasoning_content) before the final answer to improve answer quality.
Controlled by the global enable_thinking setting:
{
"enable_thinking": true
}
true: thinking is on across all channels. The Web console renders the reasoning trace; IM channels (WeChat / WeCom / DingTalk / Feishu) don't render it but still benefit from higher answer quality.false: thinking off, faster responses with lower first-token latency.temperature, top_p, presence_penalty, and frequency_penalty are silently ignored by the server (no error). CowAgent skips sending them automatically.reasoning_content on every assistant message. CowAgent handles the round-trip automatically, including across mid-session toggles of the thinking switch.Option 2: OpenAI-compatible configuration:
{
"model": "deepseek-v4-flash",
"bot_type": "openai",
"open_ai_api_key": "YOUR_API_KEY",
"open_ai_api_base": "https://api.deepseek.com/v1"
}