Open mairin opened 3 weeks ago
Hi there @mairin 👋 please have a look at the output I have provided at the PR and provide your review, if you have some time for it.
@drpaneas of course I have time - that was a record time for me to experience from filing a bug to having a fix 😁 thank you so much for the PR! I took a look and left some review comments. My main point was we support more than *.gguf format so I think that specificity can be dropped. @russellb 's comment on the logger I think is worth revisiting as well. Other than that this looks great.
Do you have some time to make the requested changes? 😃?
ready for review again ;)
Hi again, I am not working anymore on this, as per discussion it requires refactoring of input validation and debugging information (verbosity levels) in the whole project in general.
I think this issue is already solved by other commits.
(chudg1) [root@llm lab]# ilab serve --model-path models/abc.gguf --backend llama-cpp
You are using an aliased command, this will be deprecated in a future release. Please consider using `ilab model serve` instead
Failed to determine backend: Failed to determine whether the model is a GGUF format: [Errno 2] No such file or directory: 'models/abc.gguf'
(chudg1) [root@llm lab]# ls models/abc.gguf
ls: cat not access 'models/abc.gguf': No such file or directory
(chudg1) [root@llm lab]#
Describe the bug
I have models on my system that I'd already downloaded in /srv/models.... I didn't realize there were three model paths in the config.yaml to start so I got a traceback when i tried ilab serve and it couldn't find the model in the default model path.
Instead of tracing back, ideally we'd catch the error condition of the model not existing on disk and provide the user a nice human readable error message, something like:
^ something like this. @kelbrown20 prolly has better ideas on phrasing.
Here's the traceback:
Device Info (please complete the following information):