Open Adrian-1234 opened 1 year ago
Thanks for the report! I was able to reproduce the error. You can fix it by adding type=int
to the parser.add_argument
for max_tokens. I can have a PR up soon to fix this (and specify types for other arguments as well)
$python inference/bot.py --model togethercomputer/Pythia-Chat-Base-7B --max-tokens 128
Loading togethercomputer/Pythia-Chat-Base-7B to cuda:0... Loading checkpoint shards: 100%|????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????????| 2/2 [00:06<00:00, 3.26s/it] Welcome to OpenChatKit shell. Type /help or /? to list commands.