-
- [ ] [I finally got perfect labels (classification task) via prompting : r/LocalLLaMA](https://www.reddit.com/r/LocalLLaMA/comments/1amvfua/i_finally_got_perfect_labels_classification_task/)
# TIT…
-
[Edit 7/20/23]: Let's use Llama 2. AWS / Azure might have hosted versions too, so no local needed.
If there's any ticket I need engagement from the community, it's this one. Adding the ability for …
-
during the inference after the user input the model waits for few seconds but does not respond anything just returns empty. I'm using it on dell optiplex 7070 micro with intel core i7 9700t with 8 cor…
-
![image](https://github.com/AIAnytime/ChatCSV-Streamlit-App/assets/53250064/43bbbd7d-104d-4c9f-83fa-96622737984d)
-
### Required prerequisites
- [X] I have read the documentation .
- [X] I have searched the [Issue Tracker](https://github.com/XuehaiPan/nvitop/issues) that this hasn't already been reported. (comm…
-
I tested your python code simple_example.py:
python simple_example.py
File already exists in folder
llama_model_load: loading model from 'models/ggml-vicuna-13b-4bit-rev1.bin' - please wait ...
…
-
Another topic to lift from Notion to GitHub...
https://github.com/ggerganov/llama.cpp and https://github.com/ggerganov/whisper.cpp seem to have the most active development. Both are built on top of…
-
Hello,
I'm currently working on a project that requires the use of the `TheBloke/Mistral-7B-Instruct-v0.1-GGUF` model, which is in the GGUF format. I've tried using the Hugging Face library to load…
-
## Expected Behavior
I expect the UI to send my message to the selected model and show its generated answer in chat window.
## Current Behavior
It raises an error "Model not selected. Please sel…
-
Inference is painfully slow on CPU-Only setup and it seems to be because of this issue. Standard hugging face transformers do load the whole models into GPU RAM in order to maximize performance.
I'…