FMInference / FlexLLMGen

Running large language models on a single GPU for throughput-oriented scenarios.
Apache License 2.0
9.21k stars 549 forks source link

fix torchrun inference #112

Open fsx950223 opened 1 year ago