Open vitalyshalumov opened 4 months ago
We use TGI with h2oGPT just fine and we have tests.
That error is only reached if you did not set --inference_server
. So must be typo in the CLI options if you are intending to use TGI as inference server.
We use TGI with h2oGPT just fine and we have tests.
That error is only reached if you did not set
--inference_server
. So must be typo in the CLI options if you are intending to use TGI as inference server.
As I mentioned earlier, it worked on previous h2ogpt versions. In addition, it manages to generate a token using TGI at the start of the docker so typo is probably not the issue.
I understand, but the error message is literally only hit when inference_server is not set.
Thank you for your time. How does this align with " it manages to generate a token using TGI at the start of the docker"?
I have updated my h2ogpt docker version and now my docker wan't start. The error is thrown : Malformed inference server. It is a bug because few lines before in the command line it succeeded generating text with said inference server.