docs/blog/2025-04-29-db-gpt-qwen3-support.md
We are thrilled to announce that DB-GPT now supports inference with the Qwen3 series models!
Qwen3 is the latest generation of large language models in Qwen series, offering a comprehensive suite of dense and mixture-of-experts (MoE) models. Built upon extensive training, Qwen3 delivers groundbreaking advancements in reasoning, instruction-following, agent capabilities, and multilingual support, with the following key features:
Your can access the Qwen3 models according to Access to Hugging Face or ModelScope
Please read the Source Code Deployment to learn how to install DB-GPT from source code.
Qwen3 needs upgrade your transformers >= 4.51.0, please upgrade your transformers.
Here is the command to install the required dependencies for Qwen3:
# Use uv to install dependencies needed for Qwen3
# Install core dependencies and select desired extensions
uv sync --all-packages \
--extra "base" \
--extra "cuda121" \
--extra "hf" \
--extra "rag" \
--extra "storage_chromadb" \
--extra "quant_bnb" \
--extra "dbgpts" \
--extra "hf_qwen3"
To run DB-GPT with the local Qwen3 model. You can provide a configuration file to specify the model path and other parameters.
Here is an example configuration file configs/dbgpt-local-qwen3.toml:
# Model Configurations
[models]
[[models.llms]]
name = "Qwen/Qwen3-14B"
provider = "hf"
# If not provided, the model will be downloaded from the Hugging Face model hub
# uncomment the following line to specify the model path in the local file system
# path = "the-model-path-in-the-local-file-system"
[[models.embeddings]]
name = "BAAI/bge-large-zh-v1.5"
provider = "hf"
# If not provided, the model will be downloaded from the Hugging Face model hub
# uncomment the following line to specify the model path in the local file system
# path = "the-model-path-in-the-local-file-system"
In the above configuration file, [[models.llms]] specifies the LLM model, and [[models.embeddings]] specifies the embedding model. If you not provide the path parameter, the model will be downloaded from the Hugging Face model hub according to the name parameter.
Then run the following command to start the webserver:
uv run dbgpt start webserver --config configs/dbgpt-local-qwen3.toml
Open your browser and visit http://localhost:5670 to use the Qwen3 models in DB-GPT.
Enjoy the power of Qwen3 in DB-GPT!
Uniquely support of seamless switching between thinking mode (for complex logical reasoning, math, and coding) and non-thinking mode (for efficient, general-purpose dialogue) within single model, ensuring optimal performance across various scenarios.
By default, Qwen3 has thinking capabilities enabled. If you want to disable the thinking capabilities, you can set the reasoning_model=false configuration in your toml file.
[models]
[[models.llms]]
name = "Qwen/Qwen3-14B"
provider = "hf"
# Force the model to be used in non-thinking mode
reasoning_model = false
# If not provided, the model will be downloaded from the Hugging Face model hub
# uncomment the following line to specify the model path in the local file system
# path = "the-model-path-in-the-local-file-system"