Open vax-dev opened 1 year ago
@byshiue any help?
facing same issue but getting the output as [b'!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!']
Do you encounter such error by using the example used in gptj_guide.md?
yes
Sorry to barge in here, but saw these comments and I am able to replicate this behaviour on an NVIDIA GeForce GTX 1080 Ti
using the GPT J Guide
origin/dev/t5_gptj_blog
TRITON_VERSION=22.03
NVIDIA GeForce GTX 1080 Ti
11.6.1
Server Snippet:
I0209 15:56:12.313059 108 grpc_server.cc:4421] Started GRPCInferenceService at 0.0.0.0:8001 I0209 15:56:12.344709 108 http_server.cc:3113] Started HTTPService at 0.0.0.0:8000 I0209 15:56:12.387174 108 http_server.cc:178] Started Metrics Service at 0.0.0.0:8002 I0209 16:00:14.897347 108 libfastertransformer.cc:834] Start to forward I0209 16:00:14.897382 108 libfastertransformer.cc:834] Start to forward I0209 16:00:14.897539 108 libfastertransformer.cc:834] Start to forward I0209 16:00:14.897562 108 libfastertransformer.cc:834] Start to forward I0209 16:00:17.333362 108 libfastertransformer.cc:836] Stop to forward I0209 16:00:17.333413 108 libfastertransformer.cc:836] Stop to forward I0209 16:00:17.333472 108 libfastertransformer.cc:836] Stop to forward I0209 16:00:17.333564 108 libfastertransformer.cc:836] Stop to forward
Client Snippet:
Write any input prompt for the model and press ENTER: my name is [b'!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!']
BTW, Thanks for all the great work with FT backend. Cheers!
Please provide the scripts to reproduce the error. (Don't refer the guide directly because there are often some mismatch between the scripts you used and the guide).
Thanks @byshiue for your time and looking into this, I did a point by point replication of the guide Ipython Notebook to ensure I have a working setup before I bring in my scripts. Please do let me know if I am missing something here.
Updating to newer version of Faster Transformer and CUDA and building FT with lower SM 6.1 solved this issue for me. Thanks @byshiue
Hi @vax-dev,
I've been able to reproduce your setup, and I get correct results:
["to make a few comments about the book, but I will begin with an observation. The first is that I have read many books on quantum mechanics and relativity theory over the years. In fact, my dissertation was on Einstein's attempt to unify gravity and electromagnetism. My experience is that these theories are often"]
["Bob. I love him very much and he is the best thing that ever happened to me. But sometimes I get jealous of other dogs. For example, when my friend has another dog named Charlie. It makes me sad because it's not fair for me to be with someone else's dog!"]
Could you try to set the environment variable export FT_DEBUG_LEVEL=DEBUG
before starting tritonserver
? So that we can have better info what's going on on your machine.
In any case, your bad words list look wrong, i.e.:
{
"name": "bad_words_list",
"data": [[[77, 15249, 77], [2, 5, 7]]],
"dtype": "int32"
}
How did you generate it?
Description
Reproduced Steps
The output it gives is: [""]
is there any issue on it or I am running the inference in wrong way?