-
Hi! I've this error when i try to send first question. How to resolve this error?
2023-12-18 09:03:41 - Loaded .env file
2023-12-18 09:03:41 - Your app is available at http://localhost:8000
Abrie…
-
Hi all,
I'm attempting to follow the SmoothQuant tutorial for the LLAMA2-7b model: [https://github.com/intel/neural-compressor/tree/master/examples/onnxrt/nlp/huggingface_model/text_generation/llam…
-
1.执行:bash exps/finetune/mm/alpacaLlava_llamaQformerv2_13B.sh ./LLaMA2-Accessory ./13B_params.json ./tokenizer.model
报错如下:
raceback (most recent call last):
File "main_finetune.py", line 46, in
…
-
In README, there is a narrative in Performance section.
"
_-O3 includes optimizations that are expensive in terms of compile time and memory usage. Including vectorization, loop unrolling, and predi…
-
Hi, I tried to finetune Llama2-7b-chat model using megatron. I downloaded the hf checkpoint and convert it to GPT megatron checkpoint referring [https://github.com/NVIDIA/Megatron-LM/blob/fe1640a3cc48…
-
### The Feature
Could you add support for Deepspare server (see: [github](https://github.com/neuralmagic/deepsparse))? The implementation of their Quantized Pruned seems promising (see:
[sparsezoo]…
-
**Is your feature request related to a problem? Please describe.**
The code snippet we propose for llama2 models would technically "run" but the results will not be great.
**Describe the solution …
-
hello~When I run the code, this happens. Do you know why?
++++++++++++++++++++++++++++++++++50-th case++++++++++++++++++++++++++++++++++++++
+++++++++++++++++0-th question+++++++++++++++++++++
Ques…
-
Assuming that a Windows 10 machine can run this code, can this run.c lllama2 be modified to be complied to run inference mode on a Windows 10 machine?
I asked Google Gemini 1.5 Pro this question and …
-
i can not reproduce llama3-8b result according ur advice, just got
{'exact_match': 53.9604, 'num_predicted': 202, 'mean_prediction_length_characters': 1.0, 'LEval_score': 53.9604, 'display_keys': ['e…