docs/blog/2024-07-24-db-gpt-llama-3.1-support.md
We are thrilled to announce that DB-GPT now supports inference with the Meta Llama 3.1 series models!
Meta Llama 3.1 is a state-of-the-art series of language models developed by Meta AI. Designed with cutting-edge techniques, the Llama 3.1 models offer unparalleled performance and versatility. Here are some of the key highlights:
Your can access the Meta Llama 3.1 models according to Access to Hugging Face.
For comprehensive documentation and additional details, please refer to the model card.
Please read the Source Code Deployment to learn how to install DB-GPT from source code.
Llama 3.1 needs upgrade your transformers >= 4.43.0, please upgrade your transformers:
pip install --upgrade "transformers>=4.43.0"
Please cd to the DB-GPT root directory:
cd DB-GPT
We assume that your models are stored in the models directory, e.g., models/Meta-Llama-3.1-8B-Instruct.
Then modify your .env file:
LLM_MODEL=meta-llama-3.1-8b-instruct
# LLM_MODEL=meta-llama-3.1-70b-instruct
# LLM_MODEL=meta-llama-3.1-405b-instruct
## you can also specify the model path
# LLM_MODEL_PATH=models/Meta-Llama-3.1-8B-Instruct
## Quantization settings
# QUANTIZE_8bit=False
# QUANTIZE_4bit=True
## You can configure the maximum memory used by each GPU.
# MAX_GPU_MEMORY=16Gib
Then you can run the following command to start the server:
dbgpt start webserver
Open your browser and visit http://localhost:5670 to use the Meta Llama 3.1 models in DB-GPT.
Enjoy the power of Meta Llama 3.1 in DB-GPT!