-
### System Info / 系統信息
模型:glm4-9B-chat
配置文件:
data_config:
train_file: train.jsonl
val_file: dev.jsonl
test_file: dev.jsonl
num_proc: 1
max_input_length: 3500
max_output_length: 250…
-
### Description:
Some messages within the conversation encounter the following issues.
```
Diagnostic error: Cannot update a record with pending changes
```
### Environment Information…
-
Hi, We encountered a problem when using Tensorrt-LLM to infer the baichuan2-7b-chat model,Can someone please help check what's wrong。 thanks all
1. Problem: use transformers + hf model and Tensorrt…
-
Running the default example doesn't work:
```text
Namespace(verbose=True, batch_size_for_cuda_graph=1, chat_template='', model='.\\example-models\\phi2-int4-directml')
Loading model...
Model loa…
-
When converting and building llama-7b-hf with int8 kv cache and weight, there might be a memory leaking in running, each time a batch is submitted, about 2~3GB memory growing could be observed by nvid…
-
File "/usr/local/lib/python3.10/dist-packages/ragas/langchain/evalchain.py", line 166, in evaluate
dataset_with_scores = self.metric.score(dataset, callbacks=callbacks)
File "/usr/local/lib/…
-
### System Info
- `transformers` version: 4.46.2
- Platform: Linux-5.15.0-120-generic-x86_64-with-glibc2.35
- Python version: 3.10.15
- Huggingface_hub version: 0.26.2
- Safetensors version: 0.…
-
## 🐛 Bug
Hello team,
Thanks for creating such an amazing engine. I ran Llama-3-8B-Instruct-q4f16_1-MLC in server mode with different batch sizes (2-128) but I still see my requests are being run …
-
### Using a supported version?
- [X] I have searched searched open and closed issues for duplicates.
- [X] I am using Signal-Desktop as provided by the Signal team, not a 3rd-party package.
###…
-
### Checklist
- [X] I've searched for similar issues and couldn't find anything matching
- [X] I've discussed this feature request in the [OpenIMSDK Slack](https://join.slack.com/t/openimsdk/shared_i…