-
### Feature request
When using the text-generation pipeline. We would like to be able export each token as it is generated. Currently we have to wait for the generation to be completed to view the re…
-
If you attempt to run the `server.py` script outside the `text-generation-webui` directory, the `--model` argument will assume you're calling a remote model and attempts to download it from HF. Here's…
-
# Prerequisites
Please answer the following questions for yourself before submitting an issue.
- [x] I am running the latest code. Development is very rapid so there are no tagged versions as of…
-
**Describe the bug**
After installing peak through the latest statamic cli or by first installing statamic and then adding peak as a starter kit (clearing the site first), the page remains blank. The…
-
Hi,
I am trying to run the tests suite to see if my setup is correct and I am down to 31 failed, 4852 passed etc...
However the ones that failed are strange
Here a partial log, full log belo…
-
## Problem Statement
Currently, there are cases where cutting and pasting a native command line fails to run as expected in PowerShell. This may be due to incorrect parsing of quotes meant to be p…
-
I have trained (another) pruned_transducer_stateless5 model on a transcript without disfluencies and the results in offline decoding are better (decode.py of Icefall). Then, I move it to Sherpa to try…
-
I notice that you mentioned your goal of creating a drop in replacement for OpenAI. Awesome job! This is super helpful to have and especially with your demo using fastAPI.
I'm looking at langchain …
-
[GPTQ](https://arxiv.org/abs/2210.17323) is currently the SOTA one shot quantization method for LLMs.
GPTQ supports amazingly low 3-bit and 4-bit weight quantization. And it can be applied to LLaMa.
…
-
when i set --model_name_or_path llama33b-lora \
--model_name_or_path: 未找到命令