Back to Sglang

Use Models From ModelScope

docs_new/docs/supported-models/modelscope.mdx

0.5.11878 B
Original Source

To use a model from ModelScope, set the environment variable SGLANG_USE_MODELSCOPE.

bash
export SGLANG_USE_MODELSCOPE=true

We take Qwen2-7B-Instruct as an example.

Launch the Server:

bash
python -m sglang.launch_server --model-path qwen/Qwen2-7B-Instruct --port 30000

Or start it by docker:

bash
docker run --gpus all \
    -p 30000:30000 \
    -v ~/.cache/modelscope:/root/.cache/modelscope \
    --env "SGLANG_USE_MODELSCOPE=true" \
    --ipc=host \
    lmsysorg/sglang:latest \
    python3 -m sglang.launch_server --model-path Qwen/Qwen2.5-7B-Instruct --host 0.0.0.0 --port 30000

Note that modelscope uses a different cache directory than huggingface. You may need to set it manually to avoid running out of disk space.