-
Why is it said that "Please note that the RMSE and Accuracy stuffs in the training log is not reliable" for SAHP and THP ?
zhjcp updated
2 months ago
-
```
Traceback (most recent call last):
File "/root/miniconda3/envs/opensora/lib/python3.10/site-packages/gradio/queueing.py", line 541, in process_events
response = await route_utils.call_pro…
-
### System Info
- `transformers` version: 4.37.1
- Platform: Linux-4.18.0-477.27.1.el8_8.x86_64-x86_64-with-glibc2.31
- Python version: 3.10.13
- Huggingface_hub version: 0.20.3
- Safetensors ver…
-
### System Info
- `transformers` version: 4.44.0
- Platform: Linux-5.15.0-91-generic-x86_64-with-glibc2.31
- Python version: 3.10.12
- Huggingface_hub version: 0.23.4
- Safetensors version: 0.4…
-
### 🚀 The feature, motivation and pitch
Torch's embedding layers only accept int32 and int64 as input. However, for sequences with a small number of distinct possible tokens (e.g., ASCII character em…
-
**Context and question**
Getting an error in BIOMOD_ModelingOptions
**Code used**
## setup environment ----
> library(mda)
> library(gam)
> library(earth)
> library(maxnet)
> library(xg…
-
FAQ on what information modeling languages will OSIM use.
I propose we allow UML, ASN.1, and JADN, and potentially any other standard information modelling language TC Members propose.
I propose w…
-
I managed to finetune the mini-gemini mixtral model, however post finetuning I am unable to infer with the model. I tried to launch a model worker per described on the repo: `python -m minigemini.serv…
-
**Case: SQuAD task, sequence length > 512**
Does your script utilizes cached memory/extended context in a segment, such that the predictions are inferred from sequence longer than 512 tokens?
If…
-
### Describe the question.
When I run the demo of long context modeling as in https://github.com/InternLM/InternLM/tree/main/long_context
I have the following issue:
openai.OpenAIError: The api…