baichuan-inc / Baichuan-7B

A large-scale 7B pretraining language model developed by BaiChuan-Inc.
https://huggingface.co/baichuan-inc/baichuan-7B
Apache License 2.0
5.67k stars 506 forks source link

[Typo] #137

Open Chandler-Bing opened 10 months ago

Chandler-Bing commented 10 months ago

Required prerequisites

Questions

基于上述的几个优化技术,我们在千卡 A800 显卡上达到了 7B 模型 182 TFLOPS 的吞吐,GPU 峰值算力利用率高达 58.3%。

吞吐Throughput 应该指的是训练速度,eg. 3000 token/s/gpu

Checklist