docs/ascend_tutorial/examples/ascend_sglang_best_practices.rst
Last updated: 01/27/2026.
SGLang 是当前主流的高性能开源推理引擎, 昇腾已经全面原生支持该推理引擎在verl中使用, 仅需简单的构建流程,开发者即可完成环境构建,本文将提供两个经典用例来帮助开发者了解以下内容:
两个用例模型脚本以及其需要的硬件条件各自如下:
+----------------------+---------------------+----------+------------------------+
| 模型 | NPU型号 | 节点数量 | 训推后端 |
+======================+=====================+==========+========================+
| Qwen3-30B_ | Atlas 800T A3 | 1 | SGLang + Megatron |
+----------------------+---------------------+----------+------------------------+
| Qwen2.5-32B_ | Atlas 900 A2 | 2 | SGLang + FSDP |
+----------------------+---------------------+----------+------------------------+
我们在quickstart中提供了两种构建环境的方法, 1.从镜像文件DockerFile进行构建 2.从自定义Conda环境进行构建
在本实践中, 我们额外指定verl 的commit id 以避免引入其他问题
.. code-block:: bash
cd verl
git checkout c98cb8cc
1.模型数据准备
^^^^^^^^^^^
Qwen3-30B_
^^^^^^^^^^^
下载模型权重
Qwen3-30B: https://huggingface.co/Qwen/Qwen3-30B-A3B
下载数据集
DAPO-Math-17k: https://huggingface.co/datasets/BytedTsinghua-SIA/DAPO-Math-17k
HuggingFace To Megatron权重转换(可选)
.. code-block:: bash
python scripts/converter_hf_to_mcore.py
--hf_model_path Qwen/Qwen3-30B-A3B
--output_path Qwen/Qwen3-30B-A3B-mcore
--use_cpu_initialization # Only work for MoE models
注:verl当前已支持mbridge进行灵活的hf和mcore之间的权重转换,可以修改以下相关参数直接加载hf权重
.. code-block:: bash
actor_rollout_ref.actor.megatron.use_dist_checkpointing=False
actor_rollout_ref.actor.megatron.use_mbridge=True
Qwen2.5-32B_
^^^^^^^^^^^
下载模型权重
--local-dir: 模型保存路径
.. code-block:: bash
export HF_ENDPOINT=https://hf-mirror.com hf download --resume-download Qwen/Qwen2.5-32B --local-dir /path/to/local_dir
下载及处理数据集
.. code-block:: bash
wget https://huggingface.co/datasets/agentica-org/DeepScaleR-Preview-Dataset/resolve/main/deepscaler.json
python recipe/r1_ascend/json_to_parquet.py --output_dir ./data/deepscaler --json_path path/to/deepscaler.json --train_data_ratio 0.9
2.训练 ^^^^^^^^^^^ 根据开发者实际路径配置情况修改模型训练脚本中的以下参数
.. code-block:: bash
# Model Weights Paths
MODEL_PATH=Qwen/Qwen3-30B-A3B
MCORE_MODEL_PATH=Qwen/Qwen3-30B-A3B-mcore
RAY_DATA_HOME=${RAY_DATA_HOME:-"${HOME}/verl"}
CKPTS_DIR=${CKPTS_DIR:-"${RAY_DATA_HOME}/ckpts/${project_name}/${exp_name}"}
# File System Paths
TRAIN_FILE=$RAY_DATA_HOME/dataset/dapo-math-17k.parquet
TEST_FILE=$RAY_DATA_HOME/dataset/aime-2024.parquet
#保存频率,-1默认不保存,如需评测请修改此参数
trainer.save_freq=-1
对于单机任务 Qwen3-30B_ , 可以直接bash执行verl仓上示例脚本
.. code-block:: bash
bash examples/grpo_trainer/run_qwen3moe-30b_sglang_megatron_npu.sh
对于多节点任务 Qwen2.5-32B_ ,我们推荐使用以下脚本进行大规模多节点训练拉起
.. code-block:: bash
pkill -9 python ray stop --force rm -rf /tmp/ray export RAY_DEDUP_LOGS=0 export HYDRA_FULL_ERROR=1
export TASK_QUEUE_ENABLE=1 export HCCL_ASYNC_ERROR_HANDLING=0 export HCCL_EXEC_TIMEOUT=3600 export HCCL_CONNECT_TIMEOUT=3600
export HCCL_HOST_SOCKET_PORT_RANGE=60000-60050 export HCCL_NPU_SOCKET_PORT_RANGE=61000-61050 export RAY_EXPERIMENTAL_NOSET_ASCEND_RT_VISIBLE_DEVICES=1 export ASCEND_RT_VISIBLE_DEVICES=0,1,2,3,4,5,6,7,8
DEFAULT_SH="./run_*.sh" echo "Use $DEFAULT_SH"
ulimit -n 32768 mkdir logs
NNODES=2 NPUS_PER_NODE=8
MASTER_ADDR="IP FOR MASTER NODE"
SOCKET_IFNAME="Your SOCKET IFNAME" export HCCL_SOCKET_IFNAME="SOCKET IFNAME FOR CURRENT NODE" export GLOO_SOCKET_IFNAME="SOCKET IFNAME FOR CURRENT NODE"
CURRENT_IP=$(ifconfig $SOCKET_IFNAME | grep -Eo 'inet (addr:)?([0-9]{1,3}.){3}[0-9]{1,3}' | awk '{print $NF}') if [ "$MASTER_ADDR" = "$CURRENT_IP" ]; then # 主节点启动 ray start --head --port 6766 --dashboard-host=$MASTER_ADDR --node-ip-address=$CURRENT_IP --dashboard-port=8260 --resources='{"NPU": '$NPUS_PER_NODE'}'
while true; do
ray_status_output=$(ray status)
npu_count=$(echo "$ray_status_output" | grep -oP '(?<=/)\d+\.\d+(?=\s*NPU)' | head -n 1)
npu_count_int=$(echo "$npu_count" | awk '{print int($1)}')
device_count=$((npu_count_int / $NPUS_PER_NODE))
# 判断device_count 是否与 NNODES 相等
if [ "$device_count" -eq "$NNODES" ]; then
echo "Ray cluster is ready with $device_count devices (from $npu_count NPU resources), starting Python script."
ray status
bash $DEFAULT_SH
break
else
echo "Waiting for Ray to allocate $NNODES devices. Current device count: $device_count"
sleep 5
fi
done
else # 子节点尝试往主节点注册 ray 直到成功 while true; do # 尝试连接 ray 集群 ray start --address="$MASTER_ADDR:6766" --resources='{"NPU": '$NPUS_PER_NODE'}' --node-ip-address=$CURRENT_IP
# 检查连接是否成功
ray status
if [ $? -eq 0 ]; then
echo "Successfully connected to the Ray cluster!"
break
else
echo "Failed to connect to the Ray cluster. Retrying in 5 seconds..."
sleep 5
fi
done
fi
sleep 600
DEFAULT_SH:修改为训练所用配置 sh 文件路径。在此案例中修改为 Qwen2.5-32B_ 路径。
NNODES 和 NPUS_PER_NODE:修改为使用节点数量和每个节点 NPU 数量。在此案例中分别为2和8。
MASTER_ADDR:修改为对应主节点 IP。即所有节点的 MASTER_ADDR 应该相同。
SOCKET_IFNAME, HCCL_SOCKET_IFNAME, GLOO_SOCKET_IFNAME: 修改为对应通信网卡,通信网卡可以通过以下命令获取:
.. code-block:: bash
ifconfig |grep "$(hostname -I |awk '{print $1}'|awk -F '.' '{print $0}')" -B 1|awk -F ':' '{print$1}' | head -1 | tail -1
3.模型评估 ^^^^^^^^^^^
不同模型步骤一致,仅以Qwen3-30b为例列举
我们通过 AISBenchmark 评估模型,该工具支持vllm/sglang多种推理后端的评估
安装方法
.. code-block:: bash
git clone https://gitee.com/aisbench/benchmark.git cd benchmark pip install -e .
下载评估数据集
.. code-block:: bash
cd path/to/benchmark/ais_bench/datasets wget http://opencompass.oss-cn-shanghai.aliyuncs.com/datasets/data/math.zip unzip math.zip rm math.zip
修改AISBench配置代码使能sglang推理评测
打开 benchmark/ais_bench/benchmark/configs/models/vllm_api/vllm_api_stream_chat.py 文件,这是推理配置文件
.. code-block:: bash
from ais_bench.benchmark.models import VLLMCustomAPIChatStream
from ais_bench.benchmark.utils.model_postprocessors import extract_non_reasoning_content
from ais_bench.benchmark.clients import OpenAIChatStreamClient, OpenAIChatStreamSglangClient
models = [
dict(
attr="service",
type=VLLMCustomAPIChatStream,
abbr='sgl-api-stream-chat',
path="/path/to/Qwen3-30B", # 修改为 Qwen3-30B 模型路径
model="qwen3-30b",
request_rate = 0,
max_seq_len=2048,
retry = 2,
host_ip = "localhost", # 推理服务的IP
host_port = 8005, # 推理服务的端口
max_out_len = 8192, # 最大输出tokens长度
batch_size=48, # 推理的最大并发数
trust_remote_code=False,
custom_client=dict(type=OpenAIChatStreamSglangClient), #使用sglang客户端
generation_kwargs = dict(
temperature = 0,
seed = 1234,
),
pred_postprocessor=dict(type=extract_non_reasoning_content)
)
]
启动sglang_server服务
.. code-block:: bash
python -m sglang.launch_server --model-path "/path/to/Qwen3-30B" --tp-size 4 --dp-size 1 --port 8005
启动sglang_client评测
.. code-block:: bash
ais_bench --models vllm_api_stream_chat --datasets math500_gen_0_shot_cot_chat_prompt
评测结果
经过训练,模型在Math-500上的评分显著上升
+------+----------------------+---------+----------+------+----------------------+ | iter | dataset | version | metric | mode | sgl-api-stream-chat | +======+======================+=========+==========+======+======================+ | 0 | math_prm800k_500 | c4b6f0 | accuracy | gen | 84.4 | +------+----------------------+---------+----------+------+----------------------+ | 150 | math_prm800k_500 | c4b6f0 | accuracy | gen | 91.7 | +------+----------------------+---------+----------+------+----------------------+
关于NPU profiling的详细文档请参考 ascend_profiling_zh <https://github.com/volcengine/verl/blob/main/docs/ascend_tutorial/profiling/ascend_profiling_zh.rst>_
在 Qwen3-30B_ 的脚本中提供了基本的采集性能选项PROF_CONFIG,默认设置 global_profiler.steps=null 关闭采集, 开发者可根据实际需要进行参数修改
采集完成后,开发者可以使用 MindStudio Insight <https://www.hiascend.com/document/detail/zh/mindstudio/830/GUI_baseddevelopmenttool/msascendinsightug/Insight_userguide_0002.html>_ 进行数据解析
注: verl框架侧进行采集全量 Profiling 产生海量且重复的算子记录,可以根据文档修改代码仅采集关键阶段