Closed ms1design closed 6 months ago
Thanks @ms1design - just merged this, appreciate the cleanup! I noticed you added bitsandbytes
back in, which I had disabled because it no longer builds for JP6 (and was slow anyways) - did that still build for you?
And GPTQ-for-Llama
isn't really needed anymore no, I had just kept it around because it was already there for legacy purposes and still was compiling. Some of the unused/unmaintained packages I should just remove from the repo, like text-generation-inference too. Those were from when I was first exploring which inference APIs were fastest and are now just a support burden (during that time, things were also rapidly evolving with LLM/quantization APIs)
Hey @dusty-nv !
I noticed you added bitsandbytes back in, which I had disabled because it no longer builds for JP6 (and was slow anyways) - did that still build for you?
Yes, bitsandbytes
is buildable back again ;) you can find the working Dockerfile
in my PR
here: https://github.com/dusty-nv/jetson-containers/pull/420
And GPTQ-for-Llama isn't really needed anymore no...
Definitely we need a cleanup :)
Hi @dusty-nv 👋
It's a 🍒 cherry-picked PR from https://github.com/dusty-nv/jetson-containers/pull/414 to introduce improvements only for
text-generation-webui
container:torch-grammar
installation commandssentence-transformers
andflash-attention
withgit
patch (v1
&v2
supported)GPTQ-for-LLaMa
container. @dusty-nv do we still needGPTQ-for-LLaMa
as a dependency when we useauto_gptq
intext-generation-webui
container?text-generation-webui
extensions installation using nativeone_click.install_extensions_requirements
scriptsettings.json
tosettings.yaml
- when user savestext-generation-webui
UI Settings to drive, they are saved tosettings.yaml
, notsettings.json