-
### Issue Description
I am trying to run TTS, but getting `ValueError: Failed to load model from file: /Users/user/.cache/nexa/hub/official/bark-small/fp16.gguf` when I run.
### Steps to Reproduce
…
-
### Feature Description
It would be great to have a way to uninstall properly.
### Similar Features or References
For now I'm doing `nexa remove` on models downloaded.
Cleaning up `/etc/environm…
-
### multi glibc version?
[PYI-10795:ERROR] Failed to load Python shared library '/opt/nexa-linux-cuda/_internal/libpython3.11.so.1.0': dlopen: **/lib64/libm.so.6**: version `GLIBC_2.29' not found (…
-
### Issue Description
Hi there, I am seeing different results from `nexa run flux` versus `nexa run FLUX.1-schnell:q4_0`.
### Steps to Reproduce
`nexa run flux`:
```
~/repo/app/apps/web nexa ru…
-
### Question or Issue
When i download model form huggingface and Official website for the OmniVision.
I try to run the nexa server on linux, but it always response `Aborted (core dumped)`
![image]…
-
### Question and Issue
Installed Nexa SDK using option 2 : GPU ( vulcan)
Followed these instructions using $env:CMAKE_ARGS="-DGGML_VULKAN=on"; pip install nexaai --prefer-binary --index-url ht…
-
### Issue Description
$ nexa run omnivision
Downloading Octo-omni-vision/model-fp16.gguf...
An error occurred while downloading or processing the model: An unexpected error occurred: 403 Client Err…
-
### Issue Description
When I pass `--output` it to `nexa run flux`, it does not respect the output directory location.
### Steps to Reproduce
Here I try to save to `/tmp/`, but it saves to a new `g…
-
### Question or Issue
I installed the nexa by command from README:
```
CMAKE_ARGS="-DGGML_METAL=ON -DSD_METAL=ON" pip install nexaai --prefer-binary --index-url https://nexaai.github.io/nexa-sdk/…
-
Nexa devices are using 64-bit codes, but uart TX buffer is too small (48 at the moment) for sending those. Would it be too bad to handle the uart TX in a blocking way? I can make a pull request for th…