pytorch-labs / gpt-fast

Simple and efficient pytorch-native transformer text generation in <1000 LOC of python.
BSD 3-Clause "New" or "Revised" License
5.57k stars 507 forks source link

Remove unnecessary wrapper code #103

Closed HDCharles closed 7 months ago

HDCharles commented 7 months ago

Stack from ghstack (oldest at bottom):

Summary: this is inheriting from another wrapper that implements the same stuff

Test Plan:

python quantize.py --checkpoint_path checkpoints/$MODEL_REPO/model.pth --mode int4-gptq --calibration_tasks wikitext --calibration_limit 5

Reviewers:

Subscribers:

Tasks:

Tags: