-
Hi, I'm trying to run Llamaspeak following the Instructions on https://www.jetson-ai-lab.com/tutorial_llamaspeak.html
Specs:
Jetson Orin NX(16GB) Developer Kit
Jetpack 6.0 [L4T 36.3.0]
The RI…
-
Dear Greg - thanks for your continued work on this and for this follow up paper with updated pipeline. I'm the lead author of [Hermida et al. Nat Commun 2022](https://www.nature.com/articles/s41467-02…
-
### Your current environment
The output of `python collect_env.py`
```text
Collecting environment information...
/usr/local/lib/python3.10/dist-packages/vllm/connections.py:8: RuntimeWarning…
-
## ❓ Questions and Help
#### I'm trying to fine-tune BART model for classification task on another language by following the steps in the [example](https://github.com/pytorch/fairseq/blob/master/…
-
jetson@ubuntu:~$ jetson-containers run $(autotag nano_llm) python3 -m nano_llm.chat --api=mlc --model Efficient-Large-Model/VILA1.5-3b --max-context-len 256 --max-new-tokens 32 --pro…
-
## 🐛 Bug
I am not able to generate the translation with a transformer model trained according to the Iterative Product Quantization instructions (https://github.com/pytorch/fairseq/tree/master/ex…
-
Hi,
i am trying to make use of the AWQ quantization to try to load 7B LLama based models onto my RTX 3060 with 12 GB.
This fails OOM for models like https://huggingface.co/TheBloke/leo-hessianai-…
-
When I use the demo provided by read.md to run to output_ids = model.generate(**inputs, max_new_tokens=128), an error RuntimeError: Expected all tensors to be on the same device, but found at least tw…
-
Hello,
First of all great work and thanks for sharing - you are democratizing AI to newbies. I tried with my own dataset formatted like Q-A pairs and ran the generated model to infer back on trainin…
-
Hi there,
I'm not sure if this functionality already exists, but I'd like to propose a function `aggregateSamples()` which is companion to `aggregateFeatures()`, and would do essentially the same t…
csdaw updated
5 months ago