Closed leseb closed 1 week ago
Note: Links to docs will display an error until the docs builds have been completed.
There are 1 currently active SEVs. If your PR is affected, please view them below:
As of commit b41532608d0f1648285d7794eda4331b1cfb297f with merge base 46977645de6e9e29e58fada7d600c1930ed6f67b (): :green_heart: Looks good so far! There are no failures yet. :green_heart:
This comment was automatically generated by Dr. CI and updates every 15 minutes.
PR looks good, but can you share an example of when this would get triggered (i.e. when are we seeing NaN via manually kill)?
PR looks good, but can you share an example of when this would get triggered (i.e. when are we seeing NaN via manually kill)?
$ python3.10 torchchat.py chat llama3.1
NumExpr defaulting to 12 threads.
PyTorch version 2.6.0.dev20241002 available.
lm_eval is not installed, GPTQ may not be usable
Using device=mps
Loading model...
Time to load model: 15.06 seconds
-----------------------------------------------------------
Starting Interactive Chat
Entering Chat Mode. Will continue chatting back and forth with the language model until the models max context length of 8192 tokens is hit or until the user says /bye
Do you want to enter a system prompt? Enter y for yes and anything else for no.
User: /bye
Exiting Chat.
Average tokens/sec (total): nan
Average tokens/sec (first token): nan
Average tokens/sec (next tokens): nan
b4153260 fix: do not print perf stat when NaN
commit b41532608d0f1648285d7794eda4331b1cfb297f Author: Sébastien Han seb@redhat.com Date: Thu Nov 14 11:04:47 2024 +0100