-
### Search before asking
- [X] I have searched the YOLOv8 [issues](https://github.com/ultralytics/ultralytics/issues) and [discussions](https://github.com/ultralytics/ultralytics/discussions) and fou…
-
I have tried to quantize my fine-tuned llama 3 by doing these steps:
`./llama.cpp/convert.py ./Finetuned-Llama-3-70B-Instruct/ --outtype f32 --vocab-type bpe`
then :
`./llama.cpp/quantize Finetun…
-
Hi.
I'm trying to deploy [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) locally, as the documentation mentions it's supported, but it fails to generate the GGUF model file.
…
-
I have been trying to add memory to my `create_pandas_dataframe_agent` agent and ran into some issues.
I created the agent like this
```python
agent = create_pandas_dataframe_agent(
llm=llm…
-
First of all currently it is not possible to use straight `sphinx-build` command to build documentation out of source tree
```console
+ /usr/bin/sphinx-build -j48 -n -T -b man docs build/sphinx/…
-
We get a llama2-70B model, which was fine-tuned from huggingface model, and save as huggingface format in one file. We want to convert this model to meta llama2 weight . Then we can use this repo …
-
@mikecao
There is no sample application written in flight-php framework to give the developer a guideline to better coding practices using flight
I strongly suggest that a simple sample application …
-
I follow the step in the Llama smooth quantizing in this repo. I can But I got the following error message when quantize the [meta-llama/Llama-2-7b-chat-hf](https://huggingface.co/meta-llama/Llama-2-7…
-
When I use yolov3.weights to convert to onnx, and then to tensorrt, it works well.
When I use yolov3.pt(Downloaded last year)to convert to onnx, and then to tensorrt, it convers successfully, but …
-
I'm experiencing two issues below.
If there is any other way to improve the Inference speed while minimizing the loss of accuracy, please.
1. If use_quant is specified as false and executed in llm…