Back to Colossalai

README

examples/language/bert/README.md

0.5.11.3 KB
Original Source

Overview

This directory includes two parts: Using the Booster API finetune Huggingface Bert and AlBert models and benchmarking Bert and AlBert models with different Booster Plugin.

Finetune

bash test_ci.sh

Bert-Finetune Results

PluginAccuracyF1-scoreGPU number
torch_ddp84.4%88.6%2
torch_ddp_fp1684.7%88.8%2
gemini84.0%88.4%2
hybrid_parallel84.5%88.6%4

Benchmark

bash benchmark.sh

Now include these metrics in benchmark: CUDA mem occupy, throughput and the number of model parameters. If you have custom metrics, you can add them to benchmark_util.

Results

Bert

max cuda memthroughput(sample/s)params
ddp21.44 GB3.082M
ddp_fp1616.26 GB11.382M
gemini11.0 GB12.982M
low_level_zero11.29 G14.782M

AlBert

max cuda memthroughput(sample/s)params
ddpOOM
ddp_fp16OOM
gemini69.39 G1.3208M
low_level_zero56.89 G1.4208M