-
from bigdl.llm.transformers import AutoModelForCausalLM
import torch.nn.utils.prune as prune
model_path = r'D:\test_bigdl\model\Baichuan2-7B-Chat'
model = AutoModelForCausalLM.from_pretrained(mod…
-
🎉The finetuning(VQA/OCR/Grounding/Video) for Qwen2-VL-Chat series models has been supported, please check the documentation below for details:
# English
https://github.com/modelscope/ms-swift/blob/m…
-
If I understand correctly the idea should be that model generate belief states, dbsearch results, action and response conditioned on some dialog context. Then shouldn't we mask the context in between …
-
Hi, thanks for sharing the code! I am trying to reproduce the results and have some problems.
Below are my steps to train a model on the _DailyDialog_ corpus and evaluate it on the _DialSeg_711_ da…
-
- [ ] [Prompt engineering - OpenAI API](https://platform.openai.com/docs/guides/prompt-engineering/strategy-write-clear-instructions)
# Prompt Engineering - OpenAI API
## Six strategies for getting …
-
![image](https://github.com/lm-sys/FastChat/assets/131651962/4d672786-2818-4ee7-89bf-22c0ae24daae)
Using commands python3 -m fastchat.serve.cli --model-path call the dialogue robot opened by the mode…
-
```bash
2024-10-23 11:54:32,830 - _client.py[line:1038] - INFO: HTTP Request: GET http://127.0.0.1:7862/sdfiles/download?filename=&save_filename= "HTTP/1.1 200 OK"
2024-10-23 11:54:32.832 | DEBUG …
-
I try to attack llama-2-13b, and I ran the run_gcg_multiple.sh, the command is
`
export model=llama2_13
export set_name=merged
export CUDA_VISIBLE_DEVICES=7
export batch_size=512
python -…
-
- [ ] [[2303.16634] G-Eval: NLG Evaluation using GPT-4 with Better Human Alignment](https://arxiv.org/abs/2303.16634)
# [2303.16634] G-Eval: NLG Evaluation using GPT-4 with Better Human Alignment
…
-
@kadirnar
I have received the following error when running the diarization pipeline:
```python
ValueError: attempt to get argmin of an empty sequence
```
I ran the diarization example exactl…