-
### 🚀 The feature, motivation and pitch
I'm working on evaluating Llama3.1-70B on the [MMLU](https://github.com/EleutherAI/lm-evaluation-harness/blob/main/lm_eval/tasks/mmlu/README.md) and [MMLU-Pro]…
-
![motioncraft](https://github.com/user-attachments/assets/50e50a5f-90ac-4676-81db-818cb853b049)
![gt](https://github.com/user-attachments/assets/e427ba70-28a6-4886-be7d-5c183b65a365)
![paper](https:…
-
Thanks for your great work!
Would you mind sharing the code to train the evaluation model interclip? I'm investigating its performance and your help would be appreciated.
lzhyu updated
1 month ago
-
First of all, thank you so much for providing the community with an exciting open source toolkit. When I used 'UBFC-rPPG_ UBFC-rPPG_PURE_PHYSFORMER', I found that the model training results were somew…
-
### Question
I have the following problem (using the **ServiceAPI**): I have time and temperature as search parameters. When I use get_next_trial(), AxClient suggests both, which is fine for temper…
-
If I explicitly create a division by zero error:
```
[bvsm_dymola_fmi2:] Error: The following error was detected at time: 0
[bvsm_dymola_fmi2:]
Model error - division by zero: (nB*(k1*nA-k2*nC)) …
-
Not sure if this feature belongs to this library or would it require a complete separate library. I am proposing the creation of a library where llm benchmarks can be ran. For example, evaluating a mo…
-
Thank you very much for your work. I noticed that in the metric depth process, I did not evaluate.py script as I did before V1. I implemented the evaluation of kitti verification set according to the …
-
Hi,
Is is possible to provide a set of instructions to run evaluation on the validation set?
From the README:
> [test_eval.json](https://github.com/OpenDriveLab/DriveLM/blob/main/challenge…
-
When I use hpo as flaml, it can run normally according to the Quickstart code。But some errors occurred when I changed hpo to autosklearn. Is there something wrong with my code?
import pandas as pd
f…